huggingface / xet-coreLinks
xet client tech, used in huggingface_hub
☆308Updated this week
Alternatives and similar repositories for xet-core
Users that are interested in xet-core are comparing it to the libraries listed below
Sorting:
- Rust crates for XetHub☆70Updated last year
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang/tree/main/docs.☆86Updated this week
- ☆515Updated 3 weeks ago
- Super-fast Structured Outputs☆571Updated last week
- A minimalistic C++ Jinja templating engine for LLM chat templates☆193Updated last month
- ☆258Updated last week
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆210Updated 3 weeks ago
- Module, Model, and Tensor Serialization/Deserialization☆272Updated 2 months ago
- PyTorch Single Controller☆840Updated this week
- TensorRT-LLM server with Structured Outputs (JSON) built with Rust☆59Updated 6 months ago
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆108Updated 7 months ago
- 👷 Build compute kernels☆163Updated this week
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆502Updated last week
- parallel fetch☆138Updated last week
- ☆40Updated this week
- ☆449Updated this week
- Verify Precision of all Kimi K2 API Vendor☆272Updated 2 weeks ago
- Inference engine for GLiNER models, in Rust☆74Updated last week
- vLLM adapter for a TGIS-compatible gRPC server.☆42Updated this week
- Fast block-level file diffs (e.g. for VM disk images) using CoW filesystem metadata☆223Updated 4 months ago
- ☆139Updated last year
- Official Rust Implementation of Model2Vec☆139Updated last month
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- A benchmarking tool for comparing different LLM API providers' DeepSeek model deployments.☆30Updated 7 months ago
- Inference server benchmarking tool☆121Updated 3 weeks ago
- Benchmark and optimize LLM inference across frameworks with ease☆125Updated last month
- ☆134Updated last year
- Official Python API client library for turbopuffer☆83Updated last week
- Embeddable library or single binary for indexing and searching 1B vectors☆307Updated 4 months ago
- IBM development fork of https://github.com/huggingface/text-generation-inference☆61Updated last month