Tebmer / Awesome-Knowledge-Distillation-of-LLMs
This repository collects papers for "A Survey on Knowledge Distillation of Large Language Models". We break down KD into Knowledge Elicitation and Distillation Algorithms, and explore the Skill & Vertical Distillation of LLMs.
β656Updated 3 weeks ago
Related projects β
Alternatives and complementary repositories for Awesome-Knowledge-Distillation-of-LLMs
- π° Must-read papers and blogs on LLM based Long Context Modeling π₯β1,006Updated this week
- A curated list for Efficient Large Language Modelsβ1,270Updated this week
- [TMLR 2024] Efficient Large Language Models: A Surveyβ1,025Updated last week
- [ICML 2024] LESS: Selecting Influential Data for Targeted Instruction Tuningβ374Updated last month
- Awesome LLM compression research papers and tools.β1,202Updated this week
- π° Must-read papers and blogs on Speculative Decoding β‘οΈβ471Updated last week
- Reading list of hallucination in LLMs. Check out our new survey paper: "Sirenβs Song in the AI Ocean: A Survey on Hallucination in Large β¦β941Updated 2 months ago
- [NeurIPS 2023] LLM-Pruner: On the Structural Pruning of Large Language Models. Support Llama-3/3.1, Llama-2, LLaMA, BLOOM, Vicuna, Baichβ¦β874Updated last month
- [ACL 2024] LongBench: A Bilingual, Multitask Benchmark for Long Context Understandingβ671Updated 2 months ago
- [ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruningβ558Updated 8 months ago
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Rewardβ714Updated 2 weeks ago
- Fast inference from large lauguage models via speculative decodingβ569Updated 2 months ago
- O1 Replication Journey: A Strategic Progress Report β Part Iβ1,318Updated 3 weeks ago
- Codebase for Merging Language Models (ICML 2024)β774Updated 6 months ago
- β·οΈ LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-training (EMNLP 2024)β883Updated 4 months ago
- An Awesome Collection for LLM Surveyβ310Updated 2 months ago
- Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing". Your efficient and high-quality sβ¦β491Updated 2 weeks ago
- A repository sharing the literatures about long-context large language models, including the methodologies and the evaluation benchmarksβ252Updated 3 months ago
- Must-read Papers on Knowledge Editing for Large Language Models.β926Updated this week
- This is a collection of research papers for Self-Correcting Large Language Models with Automated Feedback.β436Updated 3 weeks ago
- Collection of training data management explorations for large language modelsβ286Updated 3 months ago
- A collection of AWESOME things about mixture-of-expertsβ972Updated 3 months ago
- Official Implementation of EAGLE-1 (ICML'24) and EAGLE-2 (EMNLP'24)β826Updated this week
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]β498Updated 6 months ago
- β1,271Updated this week
- LLM hallucination paper listβ292Updated 8 months ago
- A curated reading list of research in Mixture-of-Experts(MoE).β538Updated 3 weeks ago
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β647Updated last month
- AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (ICLR 2023).β275Updated last year
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ438Updated 8 months ago