Aleph-Alpha / trigrams
☆51Updated 5 months ago
Alternatives and similar repositories for trigrams:
Users that are interested in trigrams are comparing it to the libraries listed below
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆24Updated 11 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆54Updated 5 months ago
- ☆48Updated 3 months ago
- Train, tune, and infer Bamba model☆83Updated 3 weeks ago
- ☆47Updated 5 months ago
- ☆31Updated 7 months ago
- ☆71Updated 5 months ago
- LLM training in simple, raw C/CUDA☆14Updated 2 months ago
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.☆35Updated 9 months ago
- ☆42Updated last year
- A repository for research on medium sized language models.☆76Updated 8 months ago
- Experiments for efforts to train a new and improved t5☆77Updated 10 months ago
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models"☆59Updated 4 months ago
- Large language models (LLMs) made easy, EasyLM is a one stop solution for pre-training, finetuning, evaluating and serving LLMs in JAX/Fl…☆65Updated 5 months ago
- ☆41Updated 2 weeks ago
- This repo is based on https://github.com/jiaweizzhao/GaLore☆24Updated 4 months ago
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" given…☆14Updated last year
- ☆67Updated 6 months ago
- Code repository for the c-BTM paper☆105Updated last year
- ☆46Updated 8 months ago
- BPE modification that implements removing of the intermediate tokens during tokenizer training.☆25Updated 2 months ago
- ☆27Updated 3 months ago
- ☆78Updated 10 months ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆36Updated last year
- Simple replication of [ColBERT-v1](https://arxiv.org/abs/2004.12832).☆79Updated 10 months ago
- ☆25Updated last year