huggingface / xet-coreLinks
xet client tech, used in huggingface_hub
β124Updated this week
Alternatives and similar repositories for xet-core
Users that are interested in xet-core are comparing it to the libraries listed below
Sorting:
- clustering algorithm implementationβ13Updated this week
- π· Build compute kernelsβ74Updated this week
- Simple high-throughput inference libraryβ120Updated last month
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedbackβ97Updated 4 months ago
- β228Updated this week
- vLLM adapter for a TGIS-compatible gRPC server.β33Updated this week
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datasβ¦β187Updated 3 weeks ago
- TensorRT-LLM server with Structured Outputs (JSON) built with Rustβ55Updated 2 months ago
- A collection of reproducible inference engine benchmarksβ32Updated 2 months ago
- Fused Qwen3 MoE layer, compatible with HF Transformers, LoRA, 4-bit quant, Unslothβ107Updated this week
- Implementation of the paper "Lossless Compression of Vector IDs for Approximate Nearest Neighbor Search" by Severo et al.β80Updated 5 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templatesβ157Updated last month
- β29Updated 2 months ago
- OME is a Kubernetes operator for enterprise-grade management and serving of Large Language Models (LLMs)β156Updated this week
- Module, Model, and Tensor Serialization/Deserializationβ248Updated this week
- β37Updated this week
- Rust crates for XetHubβ43Updated 8 months ago
- Rust implementation of Suryaβ58Updated 4 months ago
- This repository contains statistics about the AI Infrastructure products.β18Updated 4 months ago
- Model Context Protocol Server for Apache OpenDALβ’β31Updated 3 months ago
- Benchmark for machine learning model online serving (LLM, embedding, Stable-Diffusion, Whisper)β28Updated 2 years ago
- IBM development fork of https://github.com/huggingface/text-generation-inferenceβ61Updated 2 months ago
- A Fish Speech implementation in Rust, with Candle.rsβ92Updated last month
- High-performance safetensors model loaderβ46Updated last week
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rustβ80Updated last year
- Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers, Tiktoken and more. Supports BPE, Unigram andβ¦β26Updated 3 months ago
- Inference of Mamba models in pure Cβ188Updated last year
- Simple dependency injection framework for Pythonβ21Updated last year
- β13Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language modelsβ81Updated last month