pixelspark / polyLinks
A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust
β79Updated last year
Alternatives and similar repositories for poly
Users that are interested in poly are comparing it to the libraries listed below
Sorting:
- Low rank adaptation (LoRA) for Candle.β162Updated 6 months ago
- Inference Llama 2 in one file of pure Rust π¦β233Updated 2 years ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Faceβ45Updated last year
- llm_utils: Basic LLM tools, best practices, and minimal abstraction.β47Updated 8 months ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ236Updated 2 months ago
- Andrej Karpathy's Let's build GPT: from scratch video & notebook implemented in Rust + candleβ77Updated last year
- High-level, optionally asynchronous Rust bindings to llama.cppβ232Updated last year
- Rust implementation of Suryaβ61Updated 7 months ago
- OpenAI compatible API for serving LLAMA-2 modelβ218Updated 2 years ago
- Llama2 LLM ported to Rust burnβ278Updated last year
- LLM Orchestrator built in Rustβ283Updated last year
- A collection of optimisers for use with candleβ43Updated 2 months ago
- A Fish Speech implementation in Rust, with Candle.rsβ98Updated 4 months ago
- An LLM interface (chat bot) implemented in pure Rust using HuggingFace/Candle over Axum Websockets, an SQLite Database, and a Leptos (Wasβ¦β137Updated last year
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!β110Updated 2 years ago
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python packageβ236Updated 3 weeks ago
- Library for doing RAGβ77Updated this week
- Rust port of sentence-transformers (https://github.com/UKPLab/sentence-transformers)β121Updated last year
- An extension library to Candle that provides PyTorch functions not currently available in Candleβ40Updated last year
- auto-rust is an experimental project that automatically generate Rust code with LLM (Large Language Models) during compilation, utilizingβ¦β42Updated 11 months ago
- Rust client for txtaiβ112Updated last month
- Approx nearest neighbor search in Rustβ167Updated 2 years ago
- A Rust implementation of OpenAI's Whisper model using the burn frameworkβ328Updated last year
- Extract core logic from qdrant and make it available as a library.β61Updated last year
- Stable Diffusion v1.4 ported to Rust's burn frameworkβ342Updated last year
- AI gateway and observability server written in Rust. Designed to help optimize multi-agent workflows.β64Updated last year
- Blazingly fast inference of diffusion models.β116Updated 6 months ago
- Rust bindings to https://github.com/k2-fsa/sherpa-onnxβ230Updated 2 weeks ago
- Example of tch-rs on M1β55Updated last year
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rustβ39Updated 2 years ago