Lossfunk / KernelBench-v2Links
KernelBench v2: Can LLMs Write GPU Kernels? - Benchmark with Torch -> Triton (and more!) problems
☆21Updated 5 months ago
Alternatives and similar repositories for KernelBench-v2
Users that are interested in KernelBench-v2 are comparing it to the libraries listed below
Sorting:
- The Automated LLM Speedrunning Benchmark measures how well LLM agents can reproduce previous innovations and discover new ones in languag…☆121Updated 2 months ago
- A collection of lightweight interpretability scripts to understand how LLMs think☆71Updated this week
- NanoGPT-speedrunning for the poor T4 enjoyers☆73Updated 8 months ago
- Simple repository for training small reasoning models☆47Updated 10 months ago
- LLM training in simple, raw C/CUDA☆15Updated last year
- Collection of autoregressive model implementation☆85Updated 8 months ago
- Synthetic data generation and benchmark implementation for "Episodic Memories Generation and Evaluation Benchmark for Large Language Mode…☆62Updated 2 months ago
- ☆48Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆59Updated 2 months ago
- Official repository for "BLEUBERI: BLEU is a surprisingly effective reward for instruction following"☆31Updated 6 months ago
- Official codebase for "Quantile Reward Policy Optimization: Alignment with Pointwise Regression and Exact Partition Functions" (Matrenok …☆28Updated 2 weeks ago
- ☆47Updated last year
- ☆46Updated 8 months ago
- 📄Small Batch Size Training for Language Models☆69Updated 2 months ago
- ☆91Updated last year
- Jax like function transformation engine but micro, microjax☆34Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated 2 years ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆131Updated last year
- j1-micro (1.7B) & j1-nano (600M) are absurdly tiny but mighty reward models.☆99Updated 5 months ago
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆149Updated 2 months ago
- MoE training for Me and You and maybe other people☆239Updated last week
- Train, tune, and infer Bamba model☆137Updated 6 months ago
- Official Repo for InSTA: Towards Internet-Scale Training For Agents☆55Updated 5 months ago
- an open source reproduction of NVIDIA's nGPT (Normalized Transformer with Representation Learning on the Hypersphere)☆108Updated 9 months ago
- Low memory full parameter finetuning of LLMs☆53Updated 5 months ago
- Entropy Based Sampling and Parallel CoT Decoding☆17Updated last year
- ☆28Updated last year
- An extention to the GaLore paper, to perform Natural Gradient Descent in low rank subspace☆18Updated last year
- Landing repository for the paper "Softpick: No Attention Sink, No Massive Activations with Rectified Softmax"☆85Updated 3 months ago
- ☆79Updated 2 months ago