SeanNaren / min-LLM
Minimal code to train a Large Language Model (LLM).
☆168Updated 2 years ago
Alternatives and similar repositories for min-LLM:
Users that are interested in min-LLM are comparing it to the libraries listed below
- Experiments with generating opensource language model assistants☆97Updated last year
- Multipack distributed sampler for fast padding-free training of LLMs☆186Updated 7 months ago
- Experiments on speculative sampling with Llama models☆125Updated last year
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year
- Pre-training code for Amber 7B LLM☆165Updated 10 months ago
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆110Updated last year
- Simple implementation of Speculative Sampling in NumPy for GPT-2.☆92Updated last year
- Code for the paper "The Impact of Positional Encoding on Length Generalization in Transformers", NeurIPS 2023☆130Updated 10 months ago
- Code repository for the c-BTM paper☆106Updated last year
- Pipeline for pulling and processing online language model pretraining data from the web☆175Updated last year
- batched loras☆340Updated last year
- ☆94Updated last year
- Reproduce results and replicate training fo T0 (Multitask Prompted Training Enables Zero-Shot Task Generalization)☆463Updated 2 years ago
- ☆96Updated last year
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆207Updated last year
- Landmark Attention: Random-Access Infinite Context Length for Transformers☆422Updated last year
- inference code for mixtral-8x7b-32kseqlen☆99Updated last year
- JAX implementation of the Llama 2 model☆216Updated last year
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆255Updated last year
- Tune MPTs☆84Updated last year
- Code accompanying the paper Pretraining Language Models with Human Preferences☆180Updated last year
- Simple next-token-prediction for RLHF☆222Updated last year
- Used for adaptive human in the loop evaluation of language and embedding models.☆306Updated 2 years ago
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆298Updated last year
- ☆67Updated 2 years ago
- Spherical Merge Pytorch/HF format Language Models with minimal feature loss.☆117Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆195Updated last year
- This repository contains code for extending the Stanford Alpaca synthetic instruction tuning to existing instruction-tuned models such as…☆351Updated last year
- Scaling Data-Constrained Language Models☆334Updated 5 months ago