guidance-ai / llgtrtLinks
TensorRT-LLM server with Structured Outputs (JSON) built with Rust
☆55Updated 2 months ago
Alternatives and similar repositories for llgtrt
Users that are interested in llgtrt are comparing it to the libraries listed below
Sorting:
- A high-performance constrained decoding engine based on context free grammar in Rust☆54Updated last month
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆97Updated 4 months ago
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆187Updated 3 weeks ago
- Super-fast Structured Outputs☆330Updated last week
- Official Rust Implementation of Model2Vec☆122Updated last week
- Faster structured generation☆230Updated last month
- Inference engine for GLiNER models, in Rust☆61Updated last week
- Fast serverless LLM inference, in Rust.☆88Updated 4 months ago
- ☆138Updated last year
- Super-simple, fully Rust powered "memory" (doc store + semantic search) for LLM projects, semantic search, etc.☆62Updated last year
- A simple, CUDA or CPU powered, library for creating vector embeddings using Candle and models from Hugging Face☆37Updated last year
- ☆13Updated last year
- LLama implementations benchmarking framework☆12Updated last year
- The Easiest Rust Interface for Local LLMs and an Interface for Deterministic Signals from Probabilistic LLM Vibes☆205Updated 4 months ago
- Built for demanding AI workflows, this gateway offers low-latency, provider-agnostic access, ensuring your AI applications run smoothly a…☆65Updated last month
- ☆129Updated last year
- implement llava using candle☆15Updated last year
- webgpu autograd library☆27Updated last month
- Pivotal Token Search☆109Updated last week
- xet client tech, used in huggingface_hub☆124Updated this week
- Tensor library for Zig☆11Updated 7 months ago
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆38Updated last year
- A text embedding extension for the Polars Dataframe library.☆25Updated 7 months ago
- 👷 Build compute kernels☆74Updated this week
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆89Updated 2 weeks ago
- Light WebUI for lm.rs☆24Updated 9 months ago
- ☆363Updated this week
- Experimental compiler for deep learning models☆68Updated last month
- Extracts structured data from unstructured input. Programming language agnostic. Uses llama.cpp☆45Updated last year
- Vector Database with support for late interaction and token level embeddings.☆55Updated 3 weeks ago