basetenlabs / stablelm-trussLinks
☆13Updated last year
Alternatives and similar repositories for stablelm-truss
Users that are interested in stablelm-truss are comparing it to the libraries listed below
Sorting:
- Truss for deploying Starcoder to Baseten or other platforms☆12Updated last year
- The simplest way to serve AI/ML models in production☆994Updated last week
- dstack is an open-source alternative to Kubernetes and Slurm, designed to simplify GPU allocation and AI workload orchestration for ML te…☆1,797Updated this week
- The Virtual Feature Store. Turn your existing data infrastructure into a feature store.☆1,901Updated 3 weeks ago
- A Datacenter Scale Distributed Inference Serving Framework☆4,197Updated this week
- Serving multiple LoRA finetuned LLM as one☆1,062Updated last year
- Data pipelines for AI applications☆12Updated last week
- Multi-LoRA inference server that scales to 1000s of fine-tuned LLMs☆2,996Updated 2 weeks ago
- PyTorch native quantization and sparsity for training and inference☆2,088Updated this week
- A Project Scaffolder for HUMANS.☆10Updated 7 months ago
- FlashInfer: Kernel Library for LLM Serving☆3,123Updated this week
- RayLLM - LLMs on Ray (Archived). Read README for more info.☆1,261Updated 2 months ago
- PyTriton is a Flask/FastAPI-like interface that simplifies Triton's deployment in Python environments.☆796Updated 3 months ago
- An open-source ML pipeline development platform☆988Updated 5 months ago
- S-LoRA: Serving Thousands of Concurrent LoRA Adapters☆1,830Updated last year
- Module, Model, and Tensor Serialization/Deserialization☆236Updated this week
- A throughput-oriented high-performance serving framework for LLMs☆815Updated 3 weeks ago
- Automatically Discovering Fast Parallelization Strategies for Distributed Deep Neural Network Training☆1,799Updated this week
- A fast inference library for running LLMs locally on modern consumer-class GPUs☆4,202Updated this week
- Aqueduct is no longer being maintained. Aqueduct allows you to run LLM and ML workloads on any cloud infrastructure.☆520Updated 2 years ago
- Redis for LLMs☆1,243Updated this week
- Felafax is building AI infra for non-NVIDIA GPUs☆560Updated 4 months ago
- MII makes low-latency and high-throughput inference possible, powered by DeepSpeed.☆2,019Updated 2 months ago
- Freeing data processing from scripting madness by providing a set of platform-agnostic customizable pipeline processing blocks.☆2,396Updated last week
- FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.☆836Updated 9 months ago
- The Triton TensorRT-LLM Backend☆845Updated this week
- Python client library for Modal☆350Updated this week
- ☆539Updated 7 months ago
- An ORM library for vector databases☆16Updated 2 years ago
- ⚡ Build your chatbot within minutes on your favorite device; offer SOTA compression techniques for LLMs; run LLMs efficiently on Intel Pl…☆2,170Updated 8 months ago