SakanaAI / robust-kbenchLinks
☆82Updated 2 months ago
Alternatives and similar repositories for robust-kbench
Users that are interested in robust-kbench are comparing it to the libraries listed below
Sorting:
- The evaluation framework for training-free sparse attention in LLMs☆117Updated 2 weeks ago
- Official implementation for Training LLMs with MXFP4☆118Updated 9 months ago
- Fast and memory-efficient exact attention☆75Updated 11 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆131Updated last year
- ☆131Updated 8 months ago
- Flash-Muon: An Efficient Implementation of Muon Optimizer☆233Updated 7 months ago
- Code for studying the super weight in LLM☆121Updated last year
- 🔥 LLM-powered GPU kernel synthesis: Train models to convert PyTorch ops into optimized Triton kernels via SFT+RL. Multi-turn compilation…☆116Updated 3 months ago
- ☆158Updated 11 months ago
- Kinetics: Rethinking Test-Time Scaling Laws☆86Updated 6 months ago
- ☆270Updated 8 months ago
- An efficient implementation of the NSA (Native Sparse Attention) kernel☆128Updated 7 months ago
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆92Updated 6 months ago
- Ship correct and fast LLM kernels to PyTorch☆140Updated 3 weeks ago
- Memory optimized Mixture of Experts☆73Updated 6 months ago
- Multi-Turn RL Training System with AgentTrainer for Language Model Game Reinforcement Learning☆59Updated last month
- [ICLR 2025] Palu: Compressing KV-Cache with Low-Rank Projection☆155Updated 11 months ago
- Collection of kernels written in Triton language☆178Updated 2 weeks ago
- ☆63Updated 6 months ago
- This repo contains the source code for the paper "Evolution Strategies at Scale: LLM Fine-Tuning Beyond Reinforcement Learning"☆292Updated 2 months ago
- EE-LLM is a framework for large-scale training and inference of early-exit (EE) large language models (LLMs).☆74Updated last year
- ☆105Updated 11 months ago
- ☆39Updated last year
- Physics of Language Models: Part 4.2, Canon Layers at Scale where Synthetic Pretraining Resonates in Reality☆317Updated last month
- Spectral Sphere Optimizer☆94Updated 3 weeks ago
- CodeEvolve is an open-source evolutionary coding agent for algorithm discovery and optimization.☆50Updated 3 weeks ago
- GEAR: An Efficient KV Cache Compression Recipefor Near-Lossless Generative Inference of LLM☆176Updated last year
- ☆119Updated last month
- Work in progress.☆79Updated 2 months ago
- Triton-based implementation of Sparse Mixture of Experts.☆263Updated 4 months ago