bminixhofer / tokenkitLinks
A toolkit implementing advanced methods to transfer models and model knowledge across tokenizers.
☆62Updated 7 months ago
Alternatives and similar repositories for tokenkit
Users that are interested in tokenkit are comparing it to the libraries listed below
Sorting:
- Code for Zero-Shot Tokenizer Transfer☆142Updated last year
- Official code release for "SuperBPE: Space Travel for Language Models"☆88Updated last month
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆113Updated 3 months ago
- ☆48Updated last year
- Supercharge huggingface transformers with model parallelism.☆78Updated 6 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆61Updated last year
- ☆38Updated last year
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆60Updated last year
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆94Updated last year
- Organize the Web: Constructing Domains Enhances Pre-Training Data Curation☆77Updated 9 months ago
- ☆77Updated last year
- [ICLR 2025] Monet: Mixture of Monosemantic Experts for Transformers☆75Updated 7 months ago
- code for training & evaluating Contextual Document Embedding models☆202Updated 8 months ago
- EvaByte: Efficient Byte-level Language Models at Scale☆115Updated 9 months ago
- State-of-the-art paired encoder and decoder models (17M-1B params)☆58Updated 6 months ago
- ☆59Updated last year
- ☆41Updated last year
- ☆59Updated 2 months ago
- Prune transformer layers☆74Updated last year
- [TMLR 2026] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆122Updated last year
- minimal pytorch implementation of bm25 (with sparse tensors)☆104Updated 3 months ago
- 🚢 Data Toolkit for Sailor Language Models☆95Updated 11 months ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆144Updated 2 years ago
- Official implementation of "BERTs are Generative In-Context Learners"☆32Updated 10 months ago
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆45Updated 4 months ago
- ☆26Updated last year
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆76Updated 2 weeks ago
- Language models scale reliably with over-training and on downstream tasks☆99Updated last year
- Code for the paper "Fishing for Magikarp"☆179Updated 8 months ago
- Aioli: A unified optimization framework for language model data mixing☆32Updated last year