runpod-workers / worker-vllmLinks
The RunPod worker template for serving our large language model endpoints. Powered by vLLM.
β333Updated 3 weeks ago
Alternatives and similar repositories for worker-vllm
Users that are interested in worker-vllm are comparing it to the libraries listed below
Sorting:
- A fast batching API to serve LLM modelsβ183Updated last year
- π | Python library for RunPod API and serverless worker SDK.β239Updated last week
- Examples of models deployable with Trussβ189Updated this week
- function calling-based LLM agentsβ287Updated 10 months ago
- This is our own implementation of 'Layer Selective Rank Reduction'β239Updated last year
- Low-Rank adapter extraction for fine-tuned transformers modelsβ173Updated last year
- TheBloke's Dockerfilesβ305Updated last year
- The llama-cpp-agent framework is a tool designed for easy interaction with Large Language Models (LLMs). Allowing users to chat with LLM β¦β578Updated 5 months ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAIβ222Updated last year
- One click templates for inferencing Language Modelsβ195Updated last month
- A multimodal, function calling powered LLM webui.β214Updated 9 months ago
- Tutorial for building LLM routerβ216Updated 11 months ago
- β909Updated 10 months ago
- A benchmark for emotional intelligence in large language modelsβ315Updated 11 months ago
- Merge Transformers language models by use of gradient parameters.β206Updated 11 months ago
- An OpenAI-like LLaMA inference APIβ112Updated last year
- β465Updated last year
- β157Updated last year
- A tool for generating function arguments and choosing what function to call with local LLMsβ428Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.β156Updated last year
- The easiest, and fastest way to run AI-generated Python code safelyβ327Updated 7 months ago
- β160Updated 5 months ago
- Large-scale LLM inference engineβ1,477Updated this week
- A simple Python sandbox for helpful LLM data agentsβ272Updated last year
- This code sets up a simple yet robust server using FastAPI for handling asynchronous requests for embedding generation and reranking taskβ¦β69Updated last year
- Convenient wrapper for fine-tuning and inference of Large Language Models (LLMs) with several quantization techniques (GTPQ, bitsandbytesβ¦β146Updated last year
- Convenience scripts to finetune (chat-)LLaMa3 and other models for any languageβ310Updated last year
- β199Updated last year
- β205Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.β257Updated 4 months ago