huggingface / candle-paged-attentionLinks
☆12Updated last year
Alternatives and similar repositories for candle-paged-attention
Users that are interested in candle-paged-attention are comparing it to the libraries listed below
Sorting:
- ☆19Updated last year
- A collection of optimisers for use with candle☆43Updated 3 months ago
- CLI utility to inspect and explore .safetensors and .gguf files☆34Updated 3 weeks ago
- implement llava using candle☆15Updated last year
- Experimental GPU language with meta-programming☆24Updated last year
- Rust crate for some audio utilities☆25Updated 8 months ago
- ☆13Updated 9 months ago
- ☆26Updated 7 months ago
- ☆21Updated 8 months ago
- Sample Python extension using Rust/PyO3/tch to interact with PyTorch☆38Updated last year
- Read and write tensorboard data using Rust☆23Updated last year
- Experimental compiler for deep learning models☆70Updated 2 months ago
- Graph model execution API for Candle☆16Updated 3 months ago
- GPU based FFT written in Rust and CubeCL☆24Updated 5 months ago
- 👷 Build compute kernels☆178Updated this week
- ☆17Updated last year
- Experiment of using Tangent to autodiff triton☆79Updated last year
- Your one stop CLI for ONNX model analysis.☆47Updated 3 years ago
- An implementation of the Llama architecture, to instruct and delight☆21Updated 5 months ago
- python bindings for symphonia/opus - read various audio formats from python and write opus files☆70Updated 3 months ago
- 8-bit floating point types for Rust☆60Updated 3 months ago
- ☆89Updated last year
- ☆28Updated 2 years ago
- ☆135Updated last year
- JAX bindings for Flash Attention v2☆97Updated 2 weeks ago
- Low rank adaptation (LoRA) for Candle.☆166Updated 7 months ago
- extensible collectives library in triton☆91Updated 7 months ago
- research impl of Native Sparse Attention (2502.11089)☆63Updated 9 months ago
- Simple high-throughput inference library☆149Updated 6 months ago
- A bunch of kernels that might make stuff slower 😉☆64Updated this week