pittisl / Generative-AI-TutorialLinks
A subjective learning guide for generative AI research
β84Updated last year
Alternatives and similar repositories for Generative-AI-Tutorial
Users that are interested in Generative-AI-Tutorial are comparing it to the libraries listed below
Sorting:
- β100Updated last year
- a curated list of high-quality papers on resource-efficient LLMs π±β134Updated 5 months ago
- Survey Paper List - Efficient LLM and Foundation Modelsβ255Updated 11 months ago
- [TKDE'25] The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".β412Updated last month
- A curated list of early exiting (LLM, CV, NLP, etc)β58Updated last year
- Systems for GenAIβ144Updated 4 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β110Updated last month
- Awesome-LLM-KV-Cache: A curated list of πAwesome LLM KV Cache Papers with Codes.β356Updated 6 months ago
- β56Updated 9 months ago
- Curated collection of papers in MoE model inferenceβ250Updated last month
- This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding coβ¦β187Updated last month
- Official implementation for Yuan & Liu & Zhong et al., KV Cache Compression, But What Must We Give in Return? A Comprehensive Benchmark oβ¦β82Updated 6 months ago
- [NeurIPS 2024] Implementation of paper - D-LLM: A Token Adaptive Computing Resource Allocation Strategy for Large Language Modelsβ18Updated 4 months ago
- Awesome list for LLM pruning.β255Updated last week
- A collection of 150+ surveys on LLMsβ325Updated 6 months ago
- EE-LLM is a framework for large-scale training and inference of early-exit (EE) large language models (LLMs).β67Updated last year
- EasyLiterature is an open-sourced, Python-based command line tool for automatic literature management.β288Updated last year
- This is the official Python version of CoreInfer: Accelerating Large Language Model Inference with Semantics-Inspired Adaptive Sparse Actβ¦β17Updated 10 months ago
- Reading notes on Speculative Decoding papersβ15Updated last month
- [ICLR 2025] PEARL: Parallel Speculative Decoding with Adaptive Draft Lengthβ108Updated 4 months ago
- Efficient Mixture of Experts for LLM Paper Listβ118Updated this week
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.β117Updated 3 weeks ago
- paper and its code for AI Systemβ323Updated 2 weeks ago
- [TMLR 2024] Efficient Large Language Models: A Surveyβ1,209Updated 2 months ago
- An Efficient "Factory" to Build Multiple LoRA Adaptersβ336Updated 6 months ago
- All Homeworks for TinyML and Efficient Deep Learning Computing 6.5940 β’ Fall β’ 2023 β’ https://efficientml.aiβ178Updated last year
- π° Must-read papers on KV Cache Compression (constantly updating π€).β523Updated last month
- The blog, read report and code example for AGI/LLM related knowledge.β44Updated 7 months ago
- β24Updated 2 years ago
- Pytorch implementation of our paper accepted by ICML 2024 -- CaM: Cache Merging for Memory-efficient LLMs Inferenceβ42Updated last year