coreylowman / llama-dfdxLinks
LLaMa 7b with CUDA acceleration implemented in rust. Minimal GPU memory needed!
☆110Updated 2 years ago
Alternatives and similar repositories for llama-dfdx
Users that are interested in llama-dfdx are comparing it to the libraries listed below
Sorting:
- Low rank adaptation (LoRA) for Candle.☆162Updated 6 months ago
- Llama2 LLM ported to Rust burn☆278Updated last year
- An extension library to Candle that provides PyTorch functions not currently available in Candle☆40Updated last year
- Andrej Karpathy's Let's build GPT: from scratch video & notebook implemented in Rust + candle☆77Updated last year
- Asynchronous CUDA for Rust.☆36Updated last month
- ☆93Updated 9 months ago
- A collection of optimisers for use with candle☆43Updated 2 months ago
- A minimal OpenCL, CUDA, Vulkan and host CPU array manipulation engine / framework.☆75Updated 2 months ago
- Inference Llama 2 in one file of pure Rust 🦀☆233Updated 2 years ago
- Models and examples built with Burn☆294Updated last month
- Stable Diffusion v1.4 ported to Rust's burn framework☆342Updated last year
- Example of tch-rs on M1☆55Updated last year
- Rust client for the huggingface hub aiming for minimal subset of features over `huggingface-hub` python package☆236Updated last month
- Safe, portable, high performance compute (GPGPU) kernels.☆240Updated this week
- Democratizing large model inference and training on any device.☆155Updated this week
- ☆19Updated last year
- Flexible, reusable reinforcement learning (Q learning) implementation in Rust☆158Updated last year
- A fun, hackable, GPU-accelerated, neural network library in Rust, written by an idiot☆133Updated last year
- A diffusers API in Burn (Rust)☆21Updated last year
- A neural network inference library, written in Rust.☆69Updated last year
- ☆24Updated 6 months ago
- LLaMA from First Principles☆51Updated 2 years ago
- Experimental compiler for deep learning models☆67Updated last month
- GPU based FFT written in Rust and CubeCL☆24Updated 4 months ago
- ☆58Updated 2 years ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- Fast serverless LLM inference, in Rust.☆94Updated 7 months ago
- A Discord bot, written in Rust, that generates responses using the LLaMA language model.☆95Updated 2 years ago
- 🧮 alphatensor matrix breakthrough algorithms + simd + rust.☆62Updated 3 years ago
- A Deep Learning and preprocessing framework in Rust with support for CPU and GPU.☆132Updated 2 years ago