KerfuffleV2 / llm-samplersLinks
A collection of LLM token samplers in Rust
☆18Updated last year
Alternatives and similar repositories for llm-samplers
Users that are interested in llm-samplers are comparing it to the libraries listed below
Sorting:
- A Fish Speech implementation in Rust, with Candle.rs☆98Updated 4 months ago
- High-level, optionally asynchronous Rust bindings to llama.cpp☆230Updated last year
- Low rank adaptation (LoRA) for Candle.☆162Updated 5 months ago
- ☆58Updated 2 years ago
- Rust library for whisper.cpp compatible Mel spectrograms☆75Updated 4 months ago
- A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit eva…☆94Updated 2 years ago
- A Rust implementation of OpenAI's Whisper model using the burn framework☆324Updated last year
- Rust bindings to https://github.com/k2-fsa/sherpa-onnx☆222Updated this week
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- A collection of optimisers for use with candle☆40Updated last month
- ☆32Updated 2 years ago
- An extension library to Candle that provides PyTorch functions not currently available in Candle☆39Updated last year
- ☆369Updated this week
- Stable Diffusion v1.4 ported to Rust's burn framework☆342Updated last year
- ☆20Updated last year
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python package☆232Updated last week
- Llama2 LLM ported to Rust burn☆279Updated last year
- ONNX neural network inference engine☆242Updated this week
- Rust+OpenCL+AVX2 implementation of LLaMA inference code☆548Updated last year
- GGML bindings that aim to be idiomatic Rust rather than directly corresponding to the C/C++ interface☆19Updated 2 years ago
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.☆163Updated last week
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!☆109Updated 2 years ago
- Bleeding edge low level Rust binding for GGML☆16Updated last year
- Rust port of sentence-transformers (https://github.com/UKPLab/sentence-transformers)☆121Updated last year
- Blazingly fast inference of diffusion models.☆115Updated 6 months ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆473Updated last week
- Inference Llama 2 in one file of pure Rust 🦀☆234Updated 2 years ago
- Rust bindings for OpenNMT/CTranslate2☆42Updated 2 weeks ago
- ☆24Updated 5 months ago
- Rust standalone inference of Namo-500M series models. Extremly tiny, runing VLM on CPU.☆24Updated 6 months ago