neuralmagic / vllmLinks
A high-throughput and memory-efficient inference and serving engine for LLMs
☆15Updated this week
Alternatives and similar repositories for vllm
Users that are interested in vllm are comparing it to the libraries listed below
Sorting:
- A collection of all available inference solutions for the LLMs☆91Updated 6 months ago
- ScalarLM - a unified training and inference stack☆79Updated 2 weeks ago
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Data☆42Updated this week
- ArcticInference: vLLM plugin for high-throughput, low-latency inference☆244Updated this week
- IBM development fork of https://github.com/huggingface/text-generation-inference☆61Updated last week
- ☆64Updated 6 months ago
- Nexusflow function call, tool use, and agent benchmarks.☆29Updated 9 months ago
- Framework-Agnostic RL Environments for LLM Fine-Tuning☆36Updated last week
- Accelerating your LLM training to full speed! Made with ❤️ by ServiceNow Research☆226Updated this week
- GenAI Studio is a low code platform to enable users to construct, evaluate, and benchmark GenAI applications. The platform also provide c…☆50Updated last month
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data genera…☆94Updated this week
- Cray-LM unified training and inference stack.☆22Updated 7 months ago
- Example implementation of Iteration of Tought - Gives a star if you like the project☆43Updated 9 months ago
- Simple examples using Argilla tools to build AI☆55Updated 10 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆266Updated 11 months ago
- A unified library for building, evaluating, and storing speculative decoding algorithms for LLM inference in vLLM☆48Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆89Updated last week
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆36Updated 2 months ago
- Transformer GPU VRAM estimator☆66Updated last year
- GPT-4 Level Conversational QA Trained In a Few Hours☆64Updated last year
- ☆19Updated last year
- Self-host LLMs with vLLM and BentoML☆150Updated 2 weeks ago
- Train, tune, and infer Bamba model☆132Updated 3 months ago
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆55Updated 7 months ago
- ☆18Updated last year
- ☆67Updated last year
- A simple service that integrates vLLM with Ray Serve for fast and scalable LLM serving.☆72Updated last year
- ☆68Updated 4 months ago
- 🚀 Scale your RAG pipeline using Ragswift: A scalable centralized embeddings management platform☆38Updated last year
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆132Updated this week