huggingface / xet-coreLinks
xet client tech, used in huggingface_hub
β117Updated this week
Alternatives and similar repositories for xet-core
Users that are interested in xet-core are comparing it to the libraries listed below
Sorting:
- π· Build compute kernelsβ64Updated this week
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedbackβ94Updated 3 months ago
- Rust crates for XetHubβ43Updated 8 months ago
- β28Updated 2 months ago
- TensorRT-LLM server with Structured Outputs (JSON) built with Rustβ55Updated last month
- vLLM adapter for a TGIS-compatible gRPC server.β32Updated this week
- A collection of reproducible inference engine benchmarksβ31Updated 2 months ago
- β221Updated this week
- Model Context Protocol Server for Apache OpenDALβ’β30Updated 2 months ago
- Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers, Tiktoken and more. Supports BPE, Unigram andβ¦β26Updated 3 months ago
- β13Updated last year
- clustering algorithm implementationβ13Updated last month
- A benchmarking tool for comparing different LLM API providers' DeepSeek model deployments.β29Updated 2 months ago
- Simple high-throughput inference libraryβ119Updated last month
- This repository contains statistics about the AI Infrastructure products.β18Updated 3 months ago
- β30Updated 7 months ago
- Benchmark suite for LLMs from Fireworks.aiβ76Updated 2 weeks ago
- Benchmark for machine learning model online serving (LLM, embedding, Stable-Diffusion, Whisper)β28Updated last year
- The driver for LMCache core to run in vLLMβ41Updated 4 months ago
- Rust implementation of Suryaβ58Updated 3 months ago
- β36Updated this week
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datasβ¦β177Updated this week
- β11Updated 4 months ago
- TRITONCACHE implementation of a Redis cacheβ14Updated last week
- A Fish Speech implementation in Rust, with Candle.rsβ84Updated 2 weeks ago
- Load compute kernels from the Hubβ172Updated last week
- Proof of concept for running moshi/hibiki using webrtcβ19Updated 3 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rustβ80Updated last year
- implement llava using candleβ15Updated last year
- High-performance safetensors model loaderβ39Updated 2 weeks ago