IBM / text-generation-inferenceLinks
IBM development fork of https://github.com/huggingface/text-generation-inference
☆63Updated 4 months ago
Alternatives and similar repositories for text-generation-inference
Users that are interested in text-generation-inference are comparing it to the libraries listed below
Sorting:
- Benchmark suite for LLMs from Fireworks.ai☆89Updated last week
- vLLM adapter for a TGIS-compatible gRPC server.☆50Updated this week
- 🚀 Collection of tuning recipes with HuggingFace SFTTrainer and PyTorch FSDP.☆56Updated last week
- ☆67Updated 10 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆267Updated 2 months ago
- ☆280Updated this week
- LM engine is a library for pretraining/finetuning LLMs☆113Updated this week
- ☆76Updated 7 months ago
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆138Updated last year
- Large Language Model Text Generation Inference on Habana Gaudi☆34Updated 10 months ago
- Google TPU optimizations for transformers models☆134Updated 2 weeks ago
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- experiments with inference on llama☆103Updated last year
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆131Updated 4 months ago
- Inference server benchmarking tool☆142Updated 4 months ago
- ☆56Updated last year
- A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM☆220Updated last week
- Python library for Synthetic Data Generation☆52Updated last month
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Data☆49Updated this week
- vLLM performance dashboard☆41Updated last year
- SGLang is fast serving framework for large language models and vision language models.☆32Updated 2 months ago
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated 2 weeks ago
- Hugging Face Inference Toolkit used to serve transformers, sentence-transformers, and diffusers models.☆90Updated last month
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Updated 4 months ago
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆287Updated this week
- Easy and Efficient Quantization for Transformers☆204Updated last week
- ☆44Updated last week
- ☆198Updated 2 years ago
- Train, tune, and infer Bamba model☆137Updated 8 months ago
- Module, Model, and Tensor Serialization/Deserialization☆286Updated 5 months ago