allenai / duplodocusLinks
Tooling for exact and MinHash deduplication of large-scale text datasets
☆60Updated 2 weeks ago
Alternatives and similar repositories for duplodocus
Users that are interested in duplodocus are comparing it to the libraries listed below
Sorting:
- ☆41Updated last year
- MatFormer repo☆70Updated last year
- ☆82Updated 2 months ago
- ☆48Updated last year
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆110Updated 8 months ago
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated 2 years ago
- Supercharge huggingface transformers with model parallelism.☆77Updated 6 months ago
- Data mapping framework for rust stuff☆44Updated this week
- A repository for research on medium sized language models.☆77Updated last year
- DPO, but faster 🚀☆46Updated last year
- Train, tune, and infer Bamba model☆138Updated 8 months ago
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆32Updated 7 months ago
- ☆53Updated 2 years ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆60Updated last year
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆61Updated last year
- Simple repository for training small reasoning models☆48Updated 11 months ago
- Demonstration that finetuning RoPE model on larger sequences than the pre-trained model adapts the model context limit☆63Updated 2 years ago
- ☆59Updated 2 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆23Updated last year
- Verifiers for LLM Reinforcement Learning☆80Updated 9 months ago
- ☆55Updated last year
- ☆53Updated 11 months ago
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆78Updated last year
- Simple and efficient pytorch-native transformer training and inference (batched)☆79Updated last year
- Pre-training code for CrystalCoder 7B LLM☆57Updated last year
- [TMLR 2026] When Attention Collapses: How Degenerate Layers in LLMs Enable Smaller, Stronger Models☆121Updated 11 months ago
- ☆66Updated 10 months ago
- Implementation of the paper: "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" from Google in pyTO…☆58Updated 2 weeks ago
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" given…☆15Updated 2 years ago
- vLLM adapter for a TGIS-compatible gRPC server.☆50Updated last week