Relaxed-System-Lab / COMP6211J_Course_HKUSTLinks
β42Updated 7 months ago
Alternatives and similar repositories for COMP6211J_Course_HKUST
Users that are interested in COMP6211J_Course_HKUST are comparing it to the libraries listed below
Sorting:
- Paper list for Efficient Reasoning.β548Updated 3 weeks ago
- π° Must-read papers and blogs on Speculative Decoding β‘οΈβ828Updated 3 weeks ago
- [TKDE'25] The official GitHub page for the survey paper "A Survey on Mixture of Experts in Large Language Models".β390Updated 3 weeks ago
- π° Must-read papers on KV Cache Compression (constantly updating π€).β484Updated 3 weeks ago
- Stop Overthinking: A Survey on Efficient Reasoning for Large Language Modelsβ519Updated 2 weeks ago
- Awesome-LLM-KV-Cache: A curated list of πAwesome LLM KV Cache Papers with Codes.β330Updated 4 months ago
- Survey Paper List - Efficient LLM and Foundation Modelsβ252Updated 9 months ago
- Fast inference from large lauguage models via speculative decodingβ779Updated 10 months ago
- awesome papers in LLM interpretabilityβ522Updated 3 weeks ago
- π₯ How to efficiently and effectively compress the CoTs or directly generate concise CoTs during inference while maintaining the reasoninβ¦β55Updated last month
- Awesome-Long2short-on-LRMs is a collection of state-of-the-art, novel, exciting long2short methods on large reasoning models. It containsβ¦β235Updated last month
- Official Implementation of "Learning Harmonized Representations for Speculative Sampling" (HASS)β43Updated 4 months ago
- Awesome RL Reasoning Recipes ("Triple R")β745Updated last month
- Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)β285Updated 2 months ago
- A collection of AWESOME things about mixture-of-expertsβ1,161Updated 7 months ago
- A curated reading list of research in Mixture-of-Experts(MoE).β638Updated 8 months ago
- Curated collection of papers in MoE model inferenceβ210Updated 4 months ago
- Super-Efficient RLHF Training of LLMs with Parameter Reallocationβ306Updated 2 months ago
- ReST-MCTS*: LLM Self-Training via Process Reward Guided Tree Search (NeurIPS 2024)β652Updated 5 months ago
- [NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.β459Updated 11 months ago
- [ICML 2024] Official code for the paper "Revisiting Zeroth-Order Optimization for Memory-Efficient LLM Fine-Tuning: A Benchmark ".β105Updated last week
- slime is a LLM post-training framework aiming for RL Scaling.β596Updated this week
- This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding coβ¦β164Updated 3 weeks ago
- Code associated with the paper **Draft & Verify: Lossless Large Language Model Acceleration via Self-Speculative Decoding**β194Updated 5 months ago
- β601Updated 2 months ago
- Latest Advances on System-2 Reasoningβ1,180Updated last month
- This repository contains a regularly updated paper list for LLMs-reasoning-in-latent-space.β134Updated this week
- [ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inferenceβ305Updated last week
- Official implementation of MASS: Multi-Agent Simulation Scaling for Portfolio Constructionβ139Updated last month
- β44Updated last year