facebookresearch / blt
Code for BLT research paper
☆1,513Updated this week
Alternatives and similar repositories for blt:
Users that are interested in blt are comparing it to the libraries listed below
- Training Large Language Model to Reason in a Continuous Latent Space☆1,062Updated 2 months ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆862Updated 2 months ago
- Official PyTorch implementation for "Large Language Diffusion Models"☆1,492Updated 2 weeks ago
- Recipes to scale inference-time compute of open models☆1,055Updated last month
- A Self-adaptation Framework🐙 that adapts LLMs for unseen tasks in real-time!☆1,040Updated 2 months ago
- Minimalistic large language model 3D-parallelism training☆1,793Updated this week
- NanoGPT (124M) in 3 minutes☆2,493Updated 3 weeks ago
- Minimalistic 4D-parallelism distributed training framework for education purpose☆991Updated last month
- Large Concept Models: Language modeling in a sentence representation space☆2,098Updated 2 months ago
- nanoGPT style version of Llama 3.1☆1,356Updated 8 months ago
- A bibliography and survey of the papers surrounding o1☆1,187Updated 5 months ago
- SONAR, a new multilingual and multimodal fixed-size sentence embedding space, with a full suite of speech and text encoders and decoders.☆732Updated 3 weeks ago
- [ICLR2025 Spotlight🔥] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parameters☆548Updated 2 months ago
- Unofficial implementation of Titans, SOTA memory for transformers, in Pytorch☆1,287Updated last week
- Pretraining code for a large-scale depth-recurrent language model☆743Updated last week
- 🚀 Efficient implementations of state-of-the-art linear attention models in Torch and Triton☆2,287Updated this week
- Muon optimizer: +>30% sample efficiency with <3% wallclock overhead☆575Updated 3 weeks ago
- Muon is Scalable for LLM Training☆1,022Updated 3 weeks ago
- Implementing DeepSeek R1's GRPO algorithm from scratch☆445Updated this week
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backends☆1,438Updated this week
- Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.☆1,984Updated 8 months ago
- Schedule-Free Optimization in PyTorch☆2,142Updated last week
- Helpful tools and examples for working with flex-attention☆726Updated last week
- Dream 7B, a large diffusion language model☆551Updated last week
- ☆1,015Updated 4 months ago
- Understanding R1-Zero-Like Training: A Critical Perspective☆863Updated last week
- AllenAI's post-training codebase☆2,913Updated this week
- Bringing BERT into modernity via both architecture changes and scaling☆1,322Updated 3 weeks ago
- Implementation of the sparse attention pattern proposed by the Deepseek team in their "Native Sparse Attention" paper☆593Updated 3 weeks ago
- OLMoE: Open Mixture-of-Experts Language Models☆716Updated last month