pittisl / Generative-AI-TutorialLinks
A subjective learning guide for generative AI research
β82Updated 10 months ago
Alternatives and similar repositories for Generative-AI-Tutorial
Users that are interested in Generative-AI-Tutorial are comparing it to the libraries listed below
Sorting:
- a curated list of high-quality papers on resource-efficient LLMs π±β127Updated 3 months ago
- Survey Paper List - Efficient LLM and Foundation Modelsβ248Updated 9 months ago
- β99Updated last year
- The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".β372Updated this week
- Official implementation for Yuan & Liu & Zhong et al., KV Cache Compression, But What Must We Give in Return? A Comprehensive Benchmark oβ¦β79Updated 4 months ago
- β41Updated 6 months ago
- Federated Learning Systems Paper Listβ73Updated last year
- [ICLR 2025] PEARL: Parallel Speculative Decoding with Adaptive Draft Lengthβ90Updated 2 months ago
- A curated list of early exiting (LLM, CV, NLP, etc)β55Updated 10 months ago
- Awesome list for LLM quantizationβ238Updated 2 weeks ago
- Systems for GenAIβ142Updated 2 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β105Updated 11 months ago
- All Homeworks for TinyML and Efficient Deep Learning Computing 6.5940 β’ Fall β’ 2023 β’ https://efficientml.aiβ174Updated last year
- Official Repo for "LLM-PQ: Serving LLM on Heterogeneous Clusters with Phase-Aware Partition and Adaptive Quantization"β34Updated last year
- Awesome LLM pruning papers all-in-one repository with integrating all useful resources and insights.β93Updated 6 months ago
- Awesome list for LLM pruning.β232Updated 6 months ago
- PyTorch implementation of paper "Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline".β88Updated 2 years ago
- llm theoretical performance analysis tools and support params, flops, memory and latency analysis.β96Updated 2 weeks ago
- Curated collection of papers in MoE model inferenceβ197Updated 4 months ago
- Summary of some awesome work for optimizing LLM inferenceβ77Updated 3 weeks ago
- Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)β282Updated 2 months ago
- This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding coβ¦β154Updated last week
- [ACL 2024] Not All Experts are Equal: Efficient Expert Pruning and Skipping for Mixture-of-Experts Large Language Modelsβ92Updated last year
- SQUEEZED ATTENTION: Accelerating Long Prompt LLM Inferenceβ47Updated 7 months ago
- [ASPLOS'25] Towards End-to-End Optimization of LLM-based Applications with Ayoβ25Updated last week
- [DAC 2024] EDGE-LLM: Enabling Efficient Large Language Model Adaptation on Edge Devices via Layerwise Unified Compression and Adaptive Laβ¦β57Updated 11 months ago
- Official Repo for SparseLLM: Global Pruning of LLMs (NeurIPS 2024)β62Updated 3 months ago
- β43Updated 7 months ago
- Here are my personal paper reading notes (including cloud computing, resource management, systems, machine learning, deep learning, and oβ¦β109Updated this week
- Official implementation of "Fast-dLLM: Training-free Acceleration of Diffusion LLM by Enabling KV Cache and Parallel Decoding"β233Updated 2 weeks ago