Large Language Model Text Generation Inference
☆10,854Mar 21, 2026Updated last month
Alternatives and similar repositories for text-generation-inference
Users that are interested in text-generation-inference are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- A high-throughput and memory-efficient inference and serving engine for LLMs☆79,733Updated this week
- SGLang is a high-performance serving framework for large language models and multimodal models.☆27,516Updated this week
- TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizat…☆13,604Updated this week
- A blazing fast inference solution for text embeddings models☆4,767Apr 30, 2026Updated last week
- Fast and memory-efficient exact attention☆23,736Updated this week
- Deploy on Railway without the complexity - Free Credits Offer • AdConnect your repo and Railway handles the rest with instant previews. Quickly provision container image services, databases, and storage volumes.
- Train transformer language models with reinforcement learning.☆18,349Updated this week
- An easy-to-use LLMs quantization package with user-friendly apis, based on GPTQ algorithm.☆5,059Apr 11, 2025Updated last year
- 🤗 PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.☆21,092Updated this week
- An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and Chatbot Arena.☆39,471May 1, 2026Updated last week
- Accessible large language models via k-bit quantization for PyTorch.☆8,178May 5, 2026Updated last week
- LMDeploy is a toolkit for compressing, deploying, and serving LLMs.☆7,848Updated this week
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.☆2,109Jun 30, 2025Updated 10 months ago
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆42,281Updated this week
- Transformer related optimization, including BERT, GPT☆6,415Mar 27, 2024Updated 2 years ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- LlamaIndex is the leading document agent and OCR platform☆49,354Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMs☆10,901Jun 10, 2024Updated last year
- Go ahead and axolotl questions☆11,890Updated this week
- A framework for few-shot evaluation of language models.☆12,490May 6, 2026Updated last week
- LightLLM is a Python-based LLM (Large Language Model) inference and serving framework, notable for its lightweight design, easy scalabili…☆4,046Updated this week
- Universal LLM Deployment Engine with ML Compilation☆22,598Apr 22, 2026Updated 3 weeks ago
- Ongoing research training transformer models at scale☆16,253Updated this week
- A guidance language for controlling large language models.☆21,448Updated this week
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMath☆9,482Jun 7, 2025Updated 11 months ago
- End-to-end encrypted email - Proton Mail • AdSpecial offer: 40% Off Yearly / 80% Off First Month. All Proton services are open source and independently audited for security.
- Robust recipes to align language models with human and AI preferences☆5,597Apr 8, 2026Updated last month
- Tools for merging pretrained large language models.☆7,069May 6, 2026Updated last week
- LLM inference in C/C++☆109,291Updated this week
- DSPy: The framework for programming—not prompting—language models☆34,327Updated this week
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆3,385Updated this week
- The open source codebase powering HuggingChat☆10,701Updated this week
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆9,678Updated this week
- AutoAWQ implements the AWQ algorithm for 4-bit quantization with a 2x speedup during inference. Documentation:☆2,333May 11, 2025Updated last year
- Structured Outputs☆13,825May 4, 2026Updated last week
- Deploy on Railway without the complexity - Free Credits Offer • AdConnect your repo and Railway handles the rest with instant previews. Quickly provision container image services, databases, and storage volumes.
- Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We als…☆18,318Apr 21, 2026Updated 3 weeks ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆2,918Sep 30, 2023Updated 2 years ago
- [ICLR 2024] Efficient Streaming Language Models with Attention Sinks☆7,227Jul 11, 2024Updated last year
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, Slurm, 20+ cl…☆9,953Updated this week
- The agent engineering platform. Available in TypeScript!☆136,191Updated this week
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.☆13,349May 1, 2026Updated last week
- Open-source desktop app for local LLMs. Text, vision, tool-calling, OpenAI/Anthropic-compatible API.☆46,978Updated this week