IBM / dolomite-engine
Dolomite Engine is a library for pretraining/finetuning LLMs
☆44Updated this week
Alternatives and similar repositories for dolomite-engine:
Users that are interested in dolomite-engine are comparing it to the libraries listed below
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆59Updated 5 months ago
- ☆12Updated 3 weeks ago
- Train, tune, and infer Bamba model☆86Updated 2 months ago
- A repository for research on medium sized language models.☆76Updated 10 months ago
- some common Huggingface transformers in maximal update parametrization (µP)☆80Updated 3 years ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆40Updated last year
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆71Updated 7 months ago
- ☆43Updated last year
- EvaByte: Efficient Byte-level Language Models at Scale☆85Updated last week
- ☆48Updated 4 months ago
- ☆73Updated 10 months ago
- Code for Zero-Shot Tokenizer Transfer☆125Updated 2 months ago
- The official repo for "LLoCo: Learning Long Contexts Offline"☆116Updated 9 months ago
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆58Updated 2 months ago
- Using FlexAttention to compute attention with different masking patterns☆42Updated 6 months ago
- Experiments for efforts to train a new and improved t5☆77Updated 11 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆55Updated 6 months ago
- Advanced Reasoning Benchmark Dataset for LLMs☆45Updated last year
- A toolkit for fine-tuning, inferencing, and evaluating GreenBitAI's LLMs.☆79Updated 2 weeks ago
- ☆38Updated 11 months ago
- train with kittens!☆54Updated 5 months ago
- ☆51Updated 10 months ago
- 🚀 Collection of tuning recipes with HuggingFace SFTTrainer and PyTorch FSDP.☆38Updated this week
- ArcticTraining is a framework designed to simplify and accelerate the post-training process for large language models (LLMs)☆51Updated last week
- Reference implementation for Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirectional Reward Model☆42Updated last year
- ☆48Updated last year
- A fast implementation of T5/UL2 in PyTorch using Flash Attention☆96Updated last week
- Cascade Speculative Drafting☆29Updated 11 months ago
- ☆79Updated 11 months ago
- RWKV-7: Surpassing GPT☆82Updated 4 months ago