pacman100 / LLM-Workshop
LLM Workshop by Sourab Mangrulkar
☆375Updated 10 months ago
Alternatives and similar repositories for LLM-Workshop:
Users that are interested in LLM-Workshop are comparing it to the libraries listed below
- A set of scripts and notebooks on LLM finetunning and dataset creation☆106Updated 6 months ago
- Official repository for ORPO☆448Updated 10 months ago
- ☆515Updated 5 months ago
- Implementation of paper Data Engineering for Scaling Language Models to 128K Context☆459Updated last year
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆255Updated last year
- Code for "LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding", ACL 2024☆286Updated last week
- The official evaluation suite and dynamic data release for MixEval.☆235Updated 5 months ago
- A bagel, with everything.☆320Updated last year
- Official PyTorch implementation of QA-LoRA☆131Updated last year
- Code for the paper "Rethinking Benchmark and Contamination for Language Models with Rephrased Samples"☆299Updated last year
- Manage scalable open LLM inference endpoints in Slurm clusters☆254Updated 9 months ago
- ☆529Updated 8 months ago
- Scalable toolkit for efficient model alignment☆770Updated last week
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆692Updated last year
- This project showcases an LLMOps pipeline that fine-tunes a small-size LLM model to prepare for the outage of the service LLM.☆303Updated 3 weeks ago
- Generate textbook-quality synthetic LLM pretraining data☆498Updated last year
- distributed trainer for LLMs☆573Updated 11 months ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.☆715Updated 6 months ago
- Generative Representational Instruction Tuning☆620Updated last month
- Guide for fine-tuning Llama/Mistral/CodeLlama models and more☆583Updated this week
- [ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning☆598Updated last year
- Best practices for distilling large language models.☆523Updated last year
- awesome synthetic (text) datasets☆272Updated 5 months ago
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding☆1,242Updated last month
- batched loras☆341Updated last year
- Minimalistic large language model 3D-parallelism training☆1,808Updated this week
- Automatically evaluate your LLMs in Google Colab☆615Updated 11 months ago
- experiments with inference on llama☆104Updated 10 months ago
- Easily embed, cluster and semantically label text datasets☆526Updated last year
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.☆545Updated last year