SeanNaren / min-LLMLinks
Minimal code to train a Large Language Model (LLM).
☆168Updated 2 years ago
Alternatives and similar repositories for min-LLM
Users that are interested in min-LLM are comparing it to the libraries listed below
Sorting:
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆114Updated 2 years ago
- Multipack distributed sampler for fast padding-free training of LLMs☆188Updated 9 months ago
- Code repository for the c-BTM paper☆106Updated last year
- Scalable PaLM implementation of PyTorch☆189Updated 2 years ago
- Pipeline for pulling and processing online language model pretraining data from the web☆178Updated last year
- Experiments with generating opensource language model assistants☆97Updated 2 years ago
- JAX implementation of the Llama 2 model☆217Updated last year
- Pre-training code for Amber 7B LLM☆166Updated last year
- Experiments on speculative sampling with Llama models☆126Updated last year
- Tune MPTs☆84Updated last year
- Implementation of Reinforcement Learning from Human Feedback (RLHF)☆173Updated 2 years ago
- Simple implementation of Speculative Sampling in NumPy for GPT-2.☆95Updated last year
- batched loras☆343Updated last year
- inference code for mixtral-8x7b-32kseqlen☆99Updated last year
- Scaling Data-Constrained Language Models☆334Updated 8 months ago
- A crude RLHF layer on top of nanoGPT with Gumbel-Softmax trick☆289Updated last year
- ☆67Updated 2 years ago
- GPTQ inference Triton kernel☆300Updated 2 years ago
- Used for adaptive human in the loop evaluation of language and embedding models.☆308Updated 2 years ago
- Python library which enables complex compositions of language models such as scratchpads, chain of thought, tool use, selection-inference…☆208Updated 4 months ago
- GPTQLoRA: Efficient Finetuning of Quantized LLMs with GPTQ☆103Updated 2 years ago
- An open collection of implementation tips, tricks and resources for training large language models☆473Updated 2 years ago
- ☆412Updated last year
- A puzzle to learn about prompting☆127Updated 2 years ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated 7 months ago
- TART: A plug-and-play Transformer module for task-agnostic reasoning☆196Updated last year
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆209Updated last year
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- experiments with inference on llama☆104Updated 11 months ago
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆254Updated last year