SumanthRH / tokenization
A comprehensive deep dive into the world of tokens
☆213Updated 4 months ago
Related projects ⓘ
Alternatives and complementary repositories for tokenization
- Manage scalable open LLM inference endpoints in Slurm clusters☆237Updated 3 months ago
- A bagel, with everything.☆312Updated 6 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for free☆219Updated last week
- awesome synthetic (text) datasets☆239Updated last week
- Comprehensive analysis of difference in performance of QLora, Lora, and Full Finetunes.☆81Updated last year
- ☆91Updated last year
- experiments with inference on llama☆105Updated 5 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creation☆92Updated last month
- Easily embed, cluster and semantically label text datasets☆459Updated 7 months ago
- data cleaning and curation for unstructured text☆327Updated 3 months ago
- ☆445Updated last week
- A comprehensive repository of reasoning tasks for LLMs (and beyond)☆273Updated last month
- Best practices for distilling large language models.☆392Updated 9 months ago
- code for training & evaluating Contextual Document Embedding models☆92Updated this week
- Website for hosting the Open Foundation Models Cheat Sheet.☆255Updated 4 months ago
- Fast bare-bones BPE for modern tokenizer training☆142Updated 2 weeks ago
- Multipack distributed sampler for fast padding-free training of LLMs☆175Updated 3 months ago
- Generate textbook-quality synthetic LLM pretraining data☆488Updated last year
- Let's build better datasets, together!☆202Updated 3 months ago
- ☆411Updated last year
- An Open Source Toolkit For LLM Distillation☆350Updated last month
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 lines☆194Updated 6 months ago
- batched loras☆336Updated last year
- ☆465Updated 2 months ago
- Simple Transformer in Jax☆115Updated 4 months ago
- Toolkit for attaching, training, saving and loading of new heads for transformer models☆242Updated this week
- A pipeline for LLM knowledge distillation☆77Updated 3 months ago
- A compact LLM pretrained in 9 days by using high quality data☆260Updated last month
- This is our own implementation of 'Layer Selective Rank Reduction'☆231Updated 5 months ago
- Just a bunch of benchmark logs for different LLMs☆113Updated 3 months ago