IBM / text-generation-inferenceLinks
IBM development fork of https://github.com/huggingface/text-generation-inference
☆62Updated 3 months ago
Alternatives and similar repositories for text-generation-inference
Users that are interested in text-generation-inference are comparing it to the libraries listed below
Sorting:
- Benchmark suite for LLMs from Fireworks.ai☆84Updated last month
- vLLM adapter for a TGIS-compatible gRPC server.☆47Updated this week
- 🚀 Collection of tuning recipes with HuggingFace SFTTrainer and PyTorch FSDP.☆55Updated 3 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆267Updated last month
- LM engine is a library for pretraining/finetuning LLMs☆108Updated this week
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆139Updated last year
- ☆198Updated last year
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆276Updated this week
- Google TPU optimizations for transformers models☆132Updated 3 weeks ago
- ☆138Updated 4 months ago
- ☆67Updated 9 months ago
- Data preparation code for Amber 7B LLM☆94Updated last year
- ☆75Updated 7 months ago
- 🦄 Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data …☆212Updated last week
- ☆275Updated last week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆94Updated this week
- Train, tune, and infer Bamba model☆137Updated 7 months ago
- Pre-training code for CrystalCoder 7B LLM☆56Updated last year
- A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM☆190Updated this week
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated last year
- experiments with inference on llama☆103Updated last year
- ☆43Updated this week
- Python library for Synthetic Data Generation☆51Updated 2 weeks ago
- 👷 Build compute kernels☆201Updated this week
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Updated 3 months ago
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆131Updated 3 months ago
- Module, Model, and Tensor Serialization/Deserialization☆283Updated 4 months ago
- ☆269Updated 6 months ago
- SGLang is fast serving framework for large language models and vision language models.☆31Updated last month
- ☆16Updated last month