Relaxed-System-Lab / COMP4901Y_Course_HKUST
Course Material for the UG Course COMP4901Y
β52Updated 8 months ago
Alternatives and similar repositories for COMP4901Y_Course_HKUST:
Users that are interested in COMP4901Y_Course_HKUST are comparing it to the libraries listed below
- π° Must-read papers on KV Cache Compression (constantly updating π€).β276Updated 2 weeks ago
- Systems for GenAIβ85Updated this week
- β39Updated last month
- Awesome-LLM-KV-Cache: A curated list of πAwesome LLM KV Cache Papers with Codes.β197Updated last month
- [NeurIPS 2024] Efficient LLM Scheduling by Learning to Rankβ34Updated 2 months ago
- β34Updated 2 months ago
- β62Updated 2 years ago
- [OSDI'24] Serving LLM-based Applications Efficiently with Semantic Variableβ136Updated 4 months ago
- β45Updated 3 weeks ago
- β70Updated 3 weeks ago
- A ChatGPT(GPT-3.5) & GPT-4 Workload Trace to Optimize LLM Serving Systemsβ142Updated 3 months ago
- Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)β218Updated 3 months ago
- Since the emergence of chatGPT in 2022, the acceleration of Large Language Model has become increasingly important. Here is a list of papβ¦β213Updated last month
- β17Updated 6 months ago
- β72Updated 2 years ago
- β82Updated 2 months ago
- paper and its code for AI Systemβ262Updated this week
- β40Updated last month
- A comprehensive guide for beginners in the field of data management and artificial intelligence.β142Updated 2 months ago
- PyTorch implementation of paper "Response Length Perception and Sequence Scheduling: An LLM-Empowered LLM Inference Pipeline".β81Updated last year
- Curated collection of papers in MoE model inferenceβ41Updated last week
- β84Updated 3 months ago
- Advanced Scalable Systems for Xβ29Updated last month
- The repository has collected a batch of noteworthy MLSys bloggers (Algorithms/Systems)β153Updated 3 weeks ago
- Super-Efficient RLHF Training of LLMs with Parameter Reallocationβ205Updated 2 weeks ago
- Official implementation for Yuan & Liu & Zhong et al., KV Cache Compression, But What Must We Give in Return? A Comprehensive Benchmark oβ¦β61Updated 3 weeks ago
- Survey Paper List - Efficient LLM and Foundation Modelsβ238Updated 4 months ago
- β29Updated last year
- β51Updated 10 months ago
- A Telegram bot to recommend arXiv papersβ237Updated 3 weeks ago