pixelspark / polyLinks
A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust
β80Updated last year
Alternatives and similar repositories for poly
Users that are interested in poly are comparing it to the libraries listed below
Sorting:
- Low rank adaptation (LoRA) for Candle.β152Updated 3 months ago
- Inference Llama 2 in one file of pure Rust π¦β233Updated last year
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Faceβ37Updated last year
- Rust implementation of Suryaβ58Updated 5 months ago
- Andrej Karpathy's Let's build GPT: from scratch video & notebook implemented in Rust + candleβ73Updated last year
- OpenAI compatible API for serving LLAMA-2 modelβ218Updated last year
- LLM Orchestrator built in Rustβ281Updated last year
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibesβ219Updated last week
- llm_utils: Basic LLM tools, best practices, and minimal abstraction.β46Updated 5 months ago
- A Fish Speech implementation in Rust, with Candle.rsβ94Updated last month
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!β108Updated 2 years ago
- High-level, optionally asynchronous Rust bindings to llama.cppβ226Updated last year
- Llama2 LLM ported to Rust burnβ280Updated last year
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python packageβ219Updated last month
- Rust port of sentence-transformers (https://github.com/UKPLab/sentence-transformers)β118Updated 10 months ago
- Approx nearest neighbor search in Rustβ166Updated 2 years ago
- Stable Diffusion v1.4 ported to Rust's burn frameworkβ338Updated 9 months ago
- An extension library to Candle that provides PyTorch functions not currently available in Candleβ40Updated last year
- Neural search for web-sites, docs, articles - online!β135Updated 2 months ago
- An LLM interface (chat bot) implemented in pure Rust using HuggingFace/Candle over Axum Websockets, an SQLite Database, and a Leptos (Wasβ¦β134Updated 9 months ago
- Bleeding edge low level Rust binding for GGMLβ16Updated last year
- Library for doing RAGβ74Updated last week
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.β401Updated this week
- Extract core logic from qdrant and make it available as a library.β60Updated last year
- Unofficial Rust bindings to Apple's mlx frameworkβ173Updated this week
- β31Updated 8 months ago
- Blazingly fast inference of diffusion models.β111Updated 4 months ago
- Implementing the BitNet model in Rustβ38Updated last year
- auto-rust is an experimental project that automatically generate Rust code with LLM (Large Language Models) during compilation, utilizingβ¦β40Updated 8 months ago
- Example of tch-rs on M1β54Updated last year