Antlera / nanoGPT-moe
Enable moe for nanogpt.
☆22Updated last year
Alternatives and similar repositories for nanoGPT-moe:
Users that are interested in nanoGPT-moe are comparing it to the libraries listed below
- Token Omission Via Attention☆123Updated 4 months ago
- Tree Attention: Topology-aware Decoding for Long-Context Attention on GPU clusters☆116Updated 2 months ago
- ☆66Updated 7 months ago
- ☆44Updated 3 months ago
- Repository for the paper Stream of Search: Learning to Search in Language☆138Updated 2 weeks ago
- [NAACL 2025] Official Implementation of "HMT: Hierarchical Memory Transformer for Long Context Language Processing"☆67Updated 2 weeks ago
- RWKV-7: Surpassing GPT☆79Updated 3 months ago
- Large scale 4D parallelism pre-training for 🤗 transformers in Mixture of Experts *(still work in progress)*☆81Updated last year
- NeurIPS 2024 tutorial on LLM Inference☆39Updated 2 months ago
- ☆125Updated last year
- Small and Efficient Mathematical Reasoning LLMs☆71Updated last year
- Code repository for the c-BTM paper☆105Updated last year
- This is the official repository for Inheritune.☆109Updated last week
- ☆71Updated 6 months ago
- The code repository for the CURLoRA research paper. Stable LLM continual fine-tuning and catastrophic forgetting mitigation.☆41Updated 5 months ago
- Layer-Condensed KV cache w/ 10 times larger batch size, fewer params and less computation. Dramatic speed up with better task performance…☆148Updated last month
- ☆67Updated 6 months ago
- ☆181Updated this week
- NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inference☆64Updated 2 months ago
- Code for RATIONALYST: Pre-training Process-Supervision for Improving Reasoning https://arxiv.org/pdf/2410.01044☆32Updated 4 months ago
- This is the official repository for the paper "Flora: Low-Rank Adapters Are Secretly Gradient Compressors" in ICML 2024.☆98Updated 7 months ago
- The official implementation of the paper "What Matters in Transformers? Not All Attention is Needed".☆160Updated 2 months ago
- ☆84Updated last month
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆122Updated 10 months ago
- Cold Compress is a hackable, lightweight, and open-source toolkit for creating and benchmarking cache compression methods built on top of…☆117Updated 6 months ago
- "Improving Mathematical Reasoning with Process Supervision" by OPENAI☆103Updated last week
- ☆95Updated 7 months ago
- Efficient Dictionary Learning with Switch Sparse Autoencoders (SAEs)☆20Updated 2 months ago
- ☆58Updated 9 months ago