huggingface / xet-coreLinks
xet client tech, used in huggingface_hub
☆398Updated last week
Alternatives and similar repositories for xet-core
Users that are interested in xet-core are comparing it to the libraries listed below
Sorting:
- Rust crates for XetHub☆78Updated last year
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang☆100Updated this week
- Super-fast Structured Outputs☆670Updated last week
- ☆280Updated this week
- ☆538Updated 3 months ago
- ☆44Updated this week
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆230Updated 3 weeks ago
- Module, Model, and Tensor Serialization/Deserialization☆286Updated 5 months ago
- PyTorch Single Controller☆957Updated this week
- A minimalistic C++ Jinja templating engine for LLM chat templates☆202Updated 4 months ago
- 👷 Build compute kernels☆214Updated last week
- Inference server benchmarking tool☆142Updated 4 months ago
- Real-time terminal monitor for InfiniBand networks - htop for high-speed interconnects☆134Updated last month
- Where GPUs get cooked 👩🍳🔥☆362Updated 2 weeks ago
- TensorRT-LLM server with Structured Outputs (JSON) built with Rust☆66Updated 9 months ago
- vLLM adapter for a TGIS-compatible gRPC server.☆50Updated this week
- Official Python API client library for turbopuffer☆102Updated last week
- ☆465Updated 2 months ago
- parallel fetch☆144Updated 2 months ago
- Self-hosted huggingface mirror service. 自建huggingface镜像服务。☆212Updated 6 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆307Updated 2 months ago
- GGUF implementation in C as a library and a tools CLI program☆301Updated 5 months ago
- ClearML Fractional GPU - Run multiple containers on the same GPU with driver level memory limitation ✨ and compute time-slicing☆88Updated 2 months ago
- Benchmarking suite for popular AI APIs☆88Updated last year
- Faster structured generation☆274Updated last week
- Benchmarks comparing PyTorch and MLX on Apple Silicon GPUs☆93Updated 2 weeks ago
- ☆140Updated last year
- Simple high-throughput inference library☆155Updated 8 months ago
- 🤗 Optimum ONNX: Export your model to ONNX and run inference with ONNX Runtime☆114Updated last week
- Verify Precision of all Kimi K2 API Vendor☆507Updated last week