pittisl / Generative-AI-TutorialLinks
A subjective learning guide for generative AI research
β89Updated last year
Alternatives and similar repositories for Generative-AI-Tutorial
Users that are interested in Generative-AI-Tutorial are comparing it to the libraries listed below
Sorting:
- a curated list of high-quality papers on resource-efficient LLMs π±β154Updated 9 months ago
- Survey Paper List - Efficient LLM and Foundation Modelsβ259Updated last year
- β102Updated last year
- [TKDE'25] The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".β474Updated 5 months ago
- Course materials for MIT6.5940: TinyML and Efficient Deep Learning Computingβ65Updated last year
- Systems for GenAIβ151Updated 8 months ago
- Curated collection of papers in MoE model inferenceβ329Updated 2 months ago
- [NeurIPS 2024] Implementation of paper - D-LLM: A Token Adaptive Computing Resource Allocation Strategy for Large Language Modelsβ23Updated 9 months ago
- [TMLR 2024] Efficient Large Language Models: A Surveyβ1,247Updated 6 months ago
- A curated reading list of research in Mixture-of-Experts(MoE).β656Updated last year
- Federated Learning Systems Paper Listβ75Updated last year
- [ICMLβ24] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β122Updated 6 months ago
- pytorch distribute tutorialsβ166Updated 6 months ago
- Awesome-LLM-KV-Cache: A curated list of πAwesome LLM KV Cache Papers with Codes.β405Updated 10 months ago
- Code release for book "Efficient Training in PyTorch"β118Updated 9 months ago
- All Homeworks for TinyML and Efficient Deep Learning Computing 6.5940 β’ Fall β’ 2023 β’ https://efficientml.aiβ189Updated 2 years ago
- A curated list of early exiting (LLM, CV, NLP, etc)β69Updated last year
- LLM Inference with Deep Learning Accelerator.β56Updated 11 months ago
- A collection of AWESOME things about mixture-of-expertsβ1,253Updated last year
- Awesome list for LLM pruning.β279Updated 3 months ago
- Official Repo for "SplitQuant / LLM-PQ: Resource-Efficient LLM Offline Serving on Heterogeneous GPUs via Phase-Aware Model Partition and β¦β36Updated 4 months ago
- Summary of some awesome work for optimizing LLM inferenceβ162Updated last month
- This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding coβ¦β264Updated last month
- [ICLR 2025] PEARL: Parallel Speculative Decoding with Adaptive Draft Lengthβ144Updated 2 weeks ago
- π° Must-read papers on KV Cache Compression (constantly updating π€).β631Updated 3 months ago
- β150Updated 6 months ago
- Efficient Mixture of Experts for LLM Paper Listβ154Updated 3 months ago
- Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of papβ¦β282Updated 10 months ago
- β64Updated last year
- One-size-fits-all model for mobile AI, a novel paradigm for mobile AI in which the OS and hardware co-manage a foundation model that is cβ¦β29Updated last year