IBM / dolomite-engine
Dolomite Engine is a library for pretraining/finetuning LLMs
☆21Updated this week
Related projects ⓘ
Alternatives and complementary repositories for dolomite-engine
- ☆17Updated 3 months ago
- ☆77Updated 5 months ago
- Triton Implementation of HyperAttention Algorithm☆46Updated 11 months ago
- ☆45Updated 9 months ago
- ☆71Updated 6 months ago
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆32Updated 5 months ago
- Simple and efficient pytorch-native transformer training and inference (batched)☆61Updated 7 months ago
- Awesome Triton Resources☆18Updated last month
- A MAD laboratory to improve AI architecture designs 🧪☆95Updated 6 months ago
- codebase release for EMNLP2023 paper publication☆19Updated 8 months ago
- Blog post☆16Updated 9 months ago
- Parallel Associative Scan for Language Models☆18Updated 10 months ago
- ☆38Updated 7 months ago
- A toolkit for scaling law research ⚖☆43Updated 8 months ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆56Updated 5 months ago
- ☆45Updated 4 months ago
- some common Huggingface transformers in maximal update parametrization (µP)☆76Updated 2 years ago
- Codes and files for the paper Are Emergent Abilities in Large Language Models just In-Context Learning☆34Updated 8 months ago
- The official code of "Building on Efficient Foundations: Effectively Training LLMs with Structured Feedforward Layers"☆14Updated 3 months ago
- Minimum Description Length probing for neural network representations☆16Updated last week
- Flash Attention Implementation with Multiple Backend Support and Sharding This module provides a flexible implementation of Flash Attenti…☆18Updated this week
- ☆24Updated 8 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆56Updated last month
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆35Updated 11 months ago
- ☆50Updated 6 months ago
- The simplest implementation of recent Sparse Attention patterns for efficient LLM inference.☆37Updated this week
- Official code repo for paper "Great Memory, Shallow Reasoning: Limits of kNN-LMs"☆18Updated 2 months ago
- Transformer with Mu-Parameterization, implemented in Jax/Flax. Supports FSDP on TPU pods.☆29Updated 2 weeks ago