google-deepmind / transformer_ngramsLinks
☆33Updated last year
Alternatives and similar repositories for transformer_ngrams
Users that are interested in transformer_ngrams are comparing it to the libraries listed below
Sorting:
- ☆171Updated 2 weeks ago
- Open source interpretability artefacts for R1.☆165Updated 8 months ago
- ☆162Updated 4 months ago
- ☆147Updated 3 months ago
- 📄Small Batch Size Training for Language Models☆69Updated 2 months ago
- Library for text-to-text regression, applicable to any input string representation and allows pretraining and fine-tuning over multiple r…☆301Updated 2 weeks ago
- The Automated LLM Speedrunning Benchmark measures how well LLM agents can reproduce previous innovations and discover new ones in languag…☆122Updated 2 months ago
- ☆27Updated 3 months ago
- Arrakis is a library to conduct, track and visualize mechanistic interpretability experiments.☆31Updated 8 months ago
- Curated collection of community environments☆196Updated last week
- Latent Program Network (from the "Searching Latent Program Spaces" paper)☆107Updated last month
- NanoGPT-speedrunning for the poor T4 enjoyers☆73Updated 8 months ago
- Evaluation of LLMs on latest math competitions☆205Updated last week
- ☆211Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.☆181Updated 6 months ago
- a Minimal, clean code for the Byte Pair Encoding (BPE) algorithm commonly used in LLM tokenization in pure C.☆22Updated last year
- Our solution for the arc challenge 2024☆186Updated 6 months ago
- ☆104Updated 4 months ago
- A zero-to-one guide on scaling modern transformers with n-dimensional parallelism.☆105Updated 3 months ago
- A package for defining deep learning models using categorical algebraic expressions.☆61Updated last year
- ☆116Updated 3 weeks ago
- code for training & evaluating Contextual Document Embedding models☆201Updated 7 months ago
- LLMs represent numbers on a helix and manipulate that helix to do addition.☆27Updated 10 months ago
- A collection of lightweight interpretability scripts to understand how LLMs think☆74Updated this week
- ☆29Updated last year
- $100K or 100 Days: Trade-offs when Pre-Training with Academic Resources☆149Updated 2 months ago
- Simple repository for training small reasoning models☆47Updated 10 months ago
- AlgoTune is a NeurIPS 2025 benchmark made up of 154 math, physics, and computer science problems. The goal is write code that solves each…☆79Updated this week
- Modular, scalable library to train ML models☆182Updated last week
- 🧱 Modula software package☆316Updated 4 months ago