basetenlabs / truss-examplesLinks
Examples of models deployable with Truss
β184Updated last week
Alternatives and similar repositories for truss-examples
Users that are interested in truss-examples are comparing it to the libraries listed below
Sorting:
- π | Python library for RunPod API and serverless worker SDK.β236Updated last week
- The RunPod worker template for serving our large language model endpoints. Powered by vLLM.β327Updated last week
- β115Updated 6 months ago
- A curated list of amazing RunPod projects, libraries, and resourcesβ115Updated 10 months ago
- A pipeline parallel training script for LLMs.β150Updated last month
- Gradio UI for a Cog APIβ68Updated last year
- A curated list of amazingly awesome Modal applications, demos, and shiny things. Inspired by awesome-php.β149Updated last week
- β157Updated 11 months ago
- β130Updated 2 months ago
- Starting point to build your own custom serverless endpointβ108Updated last month
- β52Updated last year
- Low-Rank adapter extraction for fine-tuned transformers modelsβ173Updated last year
- WIP - Allows you to create DSPy pipelines using ComfyUIβ189Updated 6 months ago
- π§° | RunPod CLI for pod managementβ315Updated 2 weeks ago
- Generate Synthetic Data Using OpenAI, MistralAI or AnthropicAIβ222Updated last year
- π³ | Dockerfiles for the RunPod container images used for our official templates.β193Updated this week
- automatically quant GGUF modelsβ184Updated this week
- Dagger functions to import Hugging Face GGUF models into a local ollama instance and optionally push them to ollama.com.β115Updated last year
- βοΈ | REPLACED BY https://github.com/runpod-workers | Official set of serverless worker provided by RunPod as endpoints.β59Updated 2 weeks ago
- Tutorial for building LLM routerβ211Updated 11 months ago
- The one who calls upon functions - Function-Calling Language Modelβ36Updated last year
- Create embeddings with infinity as serverless endpointβ32Updated last month
- Own your AI, search the web with itππβ86Updated 5 months ago
- A guidance compatibility layer for llama-cpp-pythonβ35Updated last year
- RunPod Serverless Worker for Oobabooga Text Generation API for LLMsβ2Updated last year
- β205Updated last year
- All the world is a play, we are but actors in it.β50Updated this week
- A Lightweight Library for AI Observabilityβ246Updated 4 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIAβs TensorRT-LLM for GPU aβ¦β43Updated 9 months ago
- Demo of AI chatbot that predicts user message to generate response quickly.β103Updated last year