octoml / triton-client-rs
A client library in Rust for Nvidia Triton.
☆27Updated last year
Alternatives and similar repositories for triton-client-rs:
Users that are interested in triton-client-rs are comparing it to the libraries listed below
- Example of tch-rs on M1☆53Updated 11 months ago
- ☆21Updated 8 months ago
- An extension library to Candle that provides PyTorch functions not currently available in Candle☆38Updated 11 months ago
- ☆28Updated 3 months ago
- Rust wrapper for Microsoft's ONNX Runtime (version 1.8)☆289Updated last year
- A collection of optimisers for use with candle☆34Updated 3 months ago
- Your one stop CLI for ONNX model analysis.☆47Updated 2 years ago
- implement llava using candle☆14Updated 9 months ago
- Asynchronous CUDA for Rust.☆30Updated 4 months ago
- A Demo server serving Bert through ONNX with GPU written in Rust with <3☆40Updated 3 years ago
- Low rank adaptation (LoRA) for Candle.☆144Updated 6 months ago
- ☆12Updated last year
- ☆19Updated 5 months ago
- A high-performance constrained decoding engine based on context free grammar in Rust☆47Updated 2 months ago
- ☆25Updated last year
- 8-bit floating point types for Rust☆46Updated last month
- A complete(grpc service and lib) Rust inference with multilingual embedding support. This version leverages the power of Rust for both GR…☆36Updated 6 months ago
- Efficent platform for inference and serving local LLMs including an OpenAI compatible API server.☆324Updated this week
- A rust port of pytorch dataloader☆26Updated 2 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- Andrej Karpathy's Let's build GPT: from scratch video & notebook implemented in Rust + candle☆71Updated 11 months ago
- Experimental compiler for deep learning models☆28Updated 3 weeks ago
- Rust wrapper for Microsoft's ONNX Runtime with CUDA support (version 1.7)☆24Updated 2 years ago
- Rust library for whisper.cpp compatible Mel spectrograms☆64Updated 2 weeks ago
- Rust implementation of Huggingface transformers pipelines using onnxruntime backend with bindings to C# and C.☆37Updated last year
- LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!☆104Updated last year
- Graph model execution API for Candle☆13Updated 3 months ago