iliao2345 / CompressARC
☆145Updated 3 weeks ago
Alternatives and similar repositories for CompressARC:
Users that are interested in CompressARC are comparing it to the libraries listed below
- Bootstrapping ARC☆115Updated 5 months ago
- Latent Program Network (from the "Searching Latent Program Spaces" paper)☆82Updated last month
- σ-GPT: A New Approach to Autoregressive Models☆63Updated 8 months ago
- Repository for the paper Stream of Search: Learning to Search in Language☆145Updated 3 months ago
- ☆237Updated 2 weeks ago
- A MAD laboratory to improve AI architecture designs 🧪☆113Updated 4 months ago
- EvaByte: Efficient Byte-level Language Models at Scale☆91Updated 2 weeks ago
- seqax = sequence modeling + JAX☆155Updated 3 weeks ago
- ☆19Updated 2 months ago
- ☆175Updated last year
- Public repository for "The Surprising Effectiveness of Test-Time Training for Abstract Reasoning"☆307Updated 5 months ago
- Efficient World Models with Context-Aware Tokenization. ICML 2024☆98Updated 7 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆105Updated this week
- 🧱 Modula software package☆188Updated last month
- ☆64Updated 10 months ago
- ☆109Updated 4 months ago
- Understand and test language model architectures on synthetic tasks.☆194Updated last month
- Normalized Transformer (nGPT)☆174Updated 5 months ago
- Minimal (400 LOC) implementation Maximum (multi-node, FSDP) GPT training☆123Updated last year
- Code to reproduce "Transformers Can Do Arithmetic with the Right Embeddings", McLeish et al (NeurIPS 2024)☆189Updated 11 months ago
- ☆94Updated 3 months ago
- ☆217Updated 9 months ago
- ☆53Updated last year
- ☆78Updated 10 months ago
- Materials for ConceptARC paper☆92Updated 5 months ago
- Benchmarking Agentic LLM and VLM Reasoning On Games☆133Updated last week
- prime-rl is a codebase for decentralized RL training at scale☆85Updated this week
- Code for NeurIPS'24 paper 'Grokked Transformers are Implicit Reasoners: A Mechanistic Journey to the Edge of Generalization'☆189Updated 5 months ago
- Cost aware hyperparameter tuning algorithm☆150Updated 10 months ago
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆123Updated 8 months ago