onehr / llama-rsLinks
Run LLaMA inference on CPU, with Rust π¦ππ¦
β28Updated 7 months ago
Alternatives and similar repositories for llama-rs
Users that are interested in llama-rs are comparing it to the libraries listed below
Sorting:
- High-level, optionally asynchronous Rust bindings to llama.cppβ228Updated last year
- Low rank adaptation (LoRA) for Candle.β154Updated 4 months ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Faceβ38Updated last year
- Andrej Karpathy's Let's build GPT: from scratch video & notebook implemented in Rust + candleβ75Updated last year
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python packageβ221Updated 2 months ago
- β350Updated this week
- Models and examples built with Burnβ278Updated last month
- Library for doing RAGβ75Updated 3 weeks ago
- ONNX neural network inference engineβ227Updated this week
- LLama.cpp rust bindingsβ398Updated last year
- Unofficial Rust bindings to Apple's mlx frameworkβ184Updated 2 weeks ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rustβ80Updated last year
- The Google mediapipe AI library. Write AI inference applications for image recognition, text classification, audio / video processing andβ¦β203Updated 11 months ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ231Updated 3 weeks ago
- Tera is an AI assistant which is tailored just for you and runs fully locally.β85Updated last year
- A Rust library integrated with ONNXRuntime, providing a collection of Computer Vison and Vision-Language models.β197Updated 3 weeks ago
- An extension library to Candle that provides PyTorch functions not currently available in Candleβ40Updated last year
- π¦ A Pure Rust Framework For Building AGI (WIP).β98Updated this week
- Use multiple LLM backends in a single crate, simple builder-based configuration, and built-in prompt chaining & templating.β135Updated 3 months ago
- Inference Llama 2 in one file of pure Rust π¦β233Updated last year
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!β108Updated 2 years ago
- A Fish Speech implementation in Rust, with Candle.rsβ94Updated 2 months ago
- A simple Rust library for OpenAI API, free from complex async operations and redundant dependencies.β193Updated last year
- An LLM interface (chat bot) implemented in pure Rust using HuggingFace/Candle over Axum Websockets, an SQLite Database, and a Leptos (Wasβ¦β136Updated 10 months ago
- β37Updated 9 months ago
- Rust bindings to https://github.com/k2-fsa/sherpa-onnxβ216Updated 3 months ago
- Rust standalone inference of Namo-500M series models. Extremly tiny, runing VLM on CPU.β24Updated 5 months ago
- Rust implementation of Ultralytics YOLOv8/v10 using ONNX (ort)β38Updated 4 months ago
- Experimental compiler for deep learning modelsβ68Updated 3 months ago
- A Pure Rust based LLM (Any LLM based MLLM such as Spark-TTS) Inference Engine, powering by Candle framework.β155Updated last month