jimexist / surya-rsLinks
Rust implementation of Surya
☆64Updated 10 months ago
Alternatives and similar repositories for surya-rs
Users that are interested in surya-rs are comparing it to the libraries listed below
Sorting:
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated 2 years ago
- A Fish Speech implementation in Rust, with Candle.rs☆106Updated 7 months ago
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆114Updated 10 months ago
- Official Rust Implementation of Model2Vec☆146Updated 3 months ago
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆88Updated 7 months ago
- LLM-as-SERP☆68Updated 10 months ago
- ☆140Updated last year
- An educational Rust project for exporting and running inference on Qwen3 LLM family☆38Updated 5 months ago
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆242Updated 5 months ago
- Fast serverless LLM inference, in Rust.☆108Updated 2 months ago
- ⚡️Lightning fast in-memory VectorDB written in rust🦀☆29Updated 10 months ago
- Kheish: A multi-role LLM agent for tasks like code auditing, file searching, and more seamlessly leveraging RAG and extensible modules.☆143Updated last year
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆65Updated 2 years ago
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆225Updated 2 weeks ago
- Run AI models anywhere. https://muna.ai/explore☆75Updated last week
- OpenAI compatible API for serving LLAMA-2 model☆218Updated 2 years ago
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆46Updated last year
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated 2 years ago
- ☆209Updated 11 months ago
- native OCR for MacOS, Windows, Linux☆201Updated 3 months ago
- Extract core logic from qdrant and make it available as a library.☆63Updated last year
- Model Context Protocol Server for Apache OpenDAL™☆34Updated 9 months ago
- llm_utils: Basic LLM tools, best practices, and minimal abstraction.☆48Updated 11 months ago
- EdgeInfer enables efficient edge intelligence by running small AI models, including embeddings and OnnxModels, on resource-constrained de…☆50Updated last year
- TensorRT-LLM server with Structured Outputs (JSON) built with Rust☆65Updated 9 months ago
- Inference engine for GLiNER models, in Rust☆83Updated 2 weeks ago
- LLM based agents with proactive interactions, long-term memory, external tool integration, and local deployment capabilities.☆107Updated 5 months ago
- The MCP enterprise actors-based server or mcp-ectors for short☆31Updated 7 months ago
- AI gateway and observability server written in Rust. Designed to help optimize multi-agent workflows.☆65Updated last year
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆39Updated 2 years ago