deepreinforce-ai / CUDA-L1Links
CUDA-L1: Improving CUDA Optimization via Contrastive Reinforcement Learning
☆238Updated 2 weeks ago
Alternatives and similar repositories for CUDA-L1
Users that are interested in CUDA-L1 are comparing it to the libraries listed below
Sorting:
- QeRL enables RL for 32B LLMs on a single H100 GPU.☆432Updated last month
- Training-free Post-training Efficient Sub-quadratic Complexity Attention. Implemented with OpenAI Triton.☆148Updated 2 weeks ago
- A collection of tricks and tools to speed up transformer models☆188Updated 3 weeks ago
- Load compute kernels from the Hub☆327Updated last week
- Efficient LLM Inference over Long Sequences☆390Updated 4 months ago
- Simple & Scalable Pretraining for Neural Architecture Research☆299Updated 2 weeks ago
- The Automated LLM Speedrunning Benchmark measures how well LLM agents can reproduce previous innovations and discover new ones in languag…☆111Updated last month
- 👷 Build compute kernels☆178Updated this week
- FlexAttention based, minimal vllm-style inference engine for fast Gemma 2 inference.☆303Updated 2 weeks ago
- Repo for "LoLCATs: On Low-Rank Linearizing of Large Language Models"☆249Updated 9 months ago
- Work in progress.☆75Updated 4 months ago
- GRadient-INformed MoE☆264Updated last year
- Official implementation of the paper: "ZClip: Adaptive Spike Mitigation for LLM Pre-Training".☆139Updated last month
- ☆254Updated 5 months ago
- PyTorch implementation of models from the Zamba2 series.☆185Updated 9 months ago
- Pytorch Distributed native training library for LLMs/VLMs with OOTB Hugging Face support☆167Updated last week
- ☆106Updated 2 weeks ago
- A safetensors extension to efficiently store sparse quantized tensors on disk☆204Updated last week
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆346Updated 6 months ago
- Official repository for the paper "NeuZip: Memory-Efficient Training and Inference with Dynamic Compression of Neural Networks". This rep…☆60Updated last year
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆130Updated 11 months ago
- ☆173Updated 4 months ago
- Samples of good AI generated CUDA kernels☆91Updated 5 months ago
- FlashRNN - Fast RNN Kernels with I/O Awareness☆163Updated 3 weeks ago
- Official PyTorch implementation for Hogwild! Inference: Parallel LLM Generation with a Concurrent Attention Cache☆129Updated 3 months ago
- Large multi-modal models (L3M) pre-training.☆222Updated last month
- ☆37Updated 5 months ago
- The evaluation framework for training-free sparse attention in LLMs☆103Updated last month
- Flash-Muon: An Efficient Implementation of Muon Optimizer☆206Updated 5 months ago
- PyTorch-native post-training at scale☆532Updated this week