Mooler0410 / LLMsPracticalGuide
A curated list of practical guide resources of LLMs (LLMs Tree, Examples, Papers)
β9,748Updated 9 months ago
Alternatives and similar repositories for LLMsPracticalGuide:
Users that are interested in LLMsPracticalGuide are comparing it to the libraries listed below
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β17,561Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ10,269Updated 8 months ago
- Train transformer language models with reinforcement learning.β12,121Updated this week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β11,401Updated 2 months ago
- The official GitHub page for the survey paper "A Survey of Large Language Models".β11,088Updated 6 months ago
- Aligning pretrained language models with instruction data generated by themselves.β4,294Updated last year
- [ICLR 2024] Fine-tuning LLaMA to follow Instructions within 1 Hour and 1.2M Parametersβ5,819Updated 11 months ago
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Adβ¦β6,036Updated 5 months ago
- 20+ high-performance LLMs with recipes to pretrain, finetune and deploy at scale.β11,691Updated last week
- Accessible large language models via k-bit quantization for PyTorch.β6,744Updated this week
- Instruct-tune LLaMA on consumer hardwareβ18,814Updated 7 months ago
- OpenLLaMA, a permissively licensed open source reproduction of Meta AIβs LLaMA 7B trained on the RedPajama datasetβ7,450Updated last year
- Awesome-LLM: a curated list of Large Language Modelβ21,784Updated last month
- Large Language Model Text Generation Inferenceβ9,833Updated this week
- Robust recipes to align language models with human and AI preferencesβ5,026Updated 3 months ago
- LLMs build upon Evol Insturct: WizardLM, WizardCoder, WizardMathβ9,348Updated 6 months ago
- Fast and memory-efficient exact attentionβ16,008Updated this week
- Instruction Tuning with GPT-4β4,275Updated last year
- Code and documentation to train Stanford's Alpaca models, and generate the data.β29,854Updated 7 months ago
- tiktoken is a fast BPE tokeniser for use with OpenAI's models.β13,581Updated 2 weeks ago
- An Extensible Toolkit for Finetuning and Inference of Large Foundation Models. Large Models for All.β8,365Updated this week
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)β¦β13,237Updated this week
- General technology for enabling AI capabilities w/ LLMs and MLLMsβ3,861Updated last month
- Numbers every LLM developer should knowβ4,185Updated last year
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)β4,588Updated last year
- A framework for few-shot evaluation of language models.β8,043Updated this week
- Universal LLM Deployment Engine with ML Compilationβ20,086Updated this week
- Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We alsβ¦β16,346Updated this week
- π A list of open LLMs available for commercial use.β11,722Updated 2 weeks ago
- The RedPajama-Data repository contains code for preparing large datasets for training large language models.β4,663Updated 2 months ago