SeanNaren / min-LLM
Minimal code to train a Large Language Model (LLM).
☆164Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for min-LLM
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆115Updated last year
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆206Updated 10 months ago
- Experiments with generating opensource language model assistants☆97Updated last year
- Pipeline for pulling and processing online language model pretraining data from the web☆174Updated last year
- ☆175Updated last year
- Code repository for the c-BTM paper☆105Updated last year
- Used for adaptive human in the loop evaluation of language and embedding models.☆303Updated last year
- Inference script for Meta's LLaMA models using Hugging Face wrapper☆111Updated last year
- DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.☆165Updated 6 months ago
- Multipack distributed sampler for fast padding-free training of LLMs☆176Updated 3 months ago
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- Scaling Data-Constrained Language Models☆321Updated last month
- Instruct-tune Open LLaMA / RedPajama / StableLM models on consumer hardware using QLoRA☆80Updated 10 months ago
- Experiments on speculative sampling with Llama models☆117Updated last year
- ☆94Updated last year
- Python library which enables complex compositions of language models such as scratchpads, chain of thought, tool use, selection-inference…☆195Updated 4 months ago
- ☆453Updated last year
- ☆158Updated last year
- ☆93Updated last year
- ☆91Updated 7 months ago
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learners☆111Updated last month
- batched loras☆336Updated last year
- A framework for few-shot evaluation of autoregressive language models.☆101Updated last year
- Finetune Falcon, LLaMA, MPT, and RedPajama on consumer hardware using PEFT LoRA☆101Updated 3 months ago
- ☆102Updated last year
- Tk-Instruct is a Transformer model that is tuned to solve many NLP tasks by following instructions.☆177Updated 2 years ago
- Implementation of Reinforcement Learning from Human Feedback (RLHF)☆169Updated last year
- Datasets collection and preprocessings framework for NLP extreme multitask learning☆149Updated 4 months ago
- Chain-of-Hindsight, A Scalable RLHF Method☆218Updated last year
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆190Updated last year