KerfuffleV2 / smolrsrwkv
A relatively basic implementation of RWKV in Rust written by someone with very little math and ML knowledge. Supports 32, 8 and 4 bit evaluation. It can also directly load PyTorch RWKV models.
☆93Updated last year
Alternatives and similar repositories for smolrsrwkv:
Users that are interested in smolrsrwkv are comparing it to the libraries listed below
- ☆32Updated last year
- GGML bindings that aim to be idiomatic Rust rather than directly corresponding to the C/C++ interface☆19Updated last year
- tinygrad port of the RWKV large language model.☆44Updated 7 months ago
- Bleeding edge low level Rust binding for GGML☆16Updated 7 months ago
- ☆57Updated last year
- Rust implementation of Huggingface transformers pipelines using onnxruntime backend with bindings to C# and C.☆36Updated last year
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!☆102Updated last year
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- Rust+OpenCL+AVX2 implementation of LLaMA inference code☆542Updated 11 months ago
- ☆25Updated last year
- High-level, optionally asynchronous Rust bindings to llama.cpp☆199Updated 7 months ago
- A collection of LLM token samplers in Rust☆16Updated last year
- 8-bit floating point types for Rust☆41Updated this week
- Implementation of the RWKV language model in pure WebGPU/Rust.☆274Updated this week
- RWKV models and examples powered by candle.☆18Updated 6 months ago
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆74Updated last year
- Experimental compiler for deep learning models☆24Updated 2 weeks ago
- Inference of Mamba models in pure C☆183Updated 11 months ago
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆37Updated last year
- Blazingly fast inference of diffusion models.☆98Updated 3 weeks ago
- GGML implementation of BERT model with Python bindings and quantization.☆53Updated 11 months ago
- ☆17Updated 3 months ago
- Tensor library for Zig☆10Updated 2 months ago
- A Fish Speech implementation in Rust, with Candle.rs☆66Updated last week
- iterate quickly with llama.cpp hot reloading. use the llama.cpp bindings with bun.sh☆47Updated last year
- Low rank adaptation (LoRA) for Candle.☆138Updated 5 months ago
- A diffusers API in Burn (Rust)☆19Updated 6 months ago
- A collection of optimisers for use with candle☆31Updated 2 months ago
- auto-rust is an experimental project that automatically generate Rust code with LLM (Large Language Models) during compilation, utilizing…☆35Updated 2 months ago
- A highly customizable, full scale web backend for web-rwkv, built on axum with websocket protocol.☆26Updated 9 months ago