jerber / lang-jepa
☆104Updated 2 months ago
Alternatives and similar repositories for lang-jepa:
Users that are interested in lang-jepa are comparing it to the libraries listed below
- smolLM with Entropix sampler on pytorch☆150Updated 4 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆168Updated 2 months ago
- Simple Transformer in Jax☆136Updated 8 months ago
- Entropy Based Sampling and Parallel CoT Decoding☆17Updated 5 months ago
- ☆124Updated this week
- ☆97Updated 5 months ago
- Draw more samples☆186Updated 8 months ago
- smol models are fun too☆89Updated 4 months ago
- Public repository for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning"☆300Updated 4 months ago
- look how they massacred my boy☆63Updated 5 months ago
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆186Updated 9 months ago
- Long context evaluation for large language models☆202Updated 2 weeks ago
- ☆80Updated 2 months ago
- A MAD laboratory to improve AI architecture designs 🧪☆107Updated 3 months ago
- The history files when recording human interaction while solving ARC tasks☆97Updated this week
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆137Updated last month
- DeMo: Decoupled Momentum Optimization☆182Updated 3 months ago
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆135Updated last week
- Repository for the paper Stream of Search: Learning to Search in Language☆142Updated last month
- Latent Program Network (from the "Searching Latent Program Spaces" paper)☆72Updated last week
- Normalized Transformer (nGPT)☆162Updated 4 months ago
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆423Updated 5 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆100Updated 4 months ago