keli-wen / AGI-StudyLinks
The blog, read report and code example for AGI/LLM related knowledge.
☆40Updated 5 months ago
Alternatives and similar repositories for AGI-Study
Users that are interested in AGI-Study are comparing it to the libraries listed below
Sorting:
- ☆140Updated last week
- slime is a LLM post-training framework aiming for RL Scaling.☆553Updated this week
- [ICLR 2025] PEARL: Parallel Speculative Decoding with Adaptive Draft Length☆93Updated 3 months ago
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆136Updated this week
- DuoDecoding: Hardware-aware Heterogeneous Speculative Decoding with Dynamic Multi-Sequence Drafting☆15Updated 4 months ago
- a survey of long-context LLMs from four perspectives, architecture, infrastructure, training, and evaluation☆52Updated 3 months ago
- ☆193Updated 3 months ago
- ☆142Updated 4 months ago
- [ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference☆303Updated this week
- ☆47Updated 2 weeks ago
- Multi-Candidate Speculative Decoding☆35Updated last year
- qwen-nsa☆68Updated 3 months ago
- This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding co…☆162Updated 3 weeks ago
- Code associated with the paper **Draft & Verify: Lossless Large Language Model Acceleration via Self-Speculative Decoding**☆194Updated 5 months ago
- Pretrain、decay、SFT a CodeLLM from scratch 🧙♂️☆36Updated last year
- ZO2 (Zeroth-Order Offloading): Full Parameter Fine-Tuning 175B LLMs with 18GB GPU Memory☆148Updated 3 weeks ago
- A Comprehensive Survey on Long Context Language Modeling☆161Updated this week
- Official implementation of "Fast-dLLM: Training-free Acceleration of Diffusion LLM by Enabling KV Cache and Parallel Decoding"☆278Updated last week
- VeOmni: Scaling any Modality Model Training to any Accelerators with PyTorch native Training Framework☆370Updated this week
- Efficient Mixture of Experts for LLM Paper List☆79Updated 6 months ago
- Due to the huge vocaburary size (151,936) of Qwen models, the Embedding and LM Head weights are excessively heavy. Therefore, this projec…☆23Updated 10 months ago
- Reproducing R1 for Code with Reliable Rewards☆232Updated 2 months ago
- Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)☆285Updated 2 months ago
- ☆202Updated 8 months ago
- ☆261Updated last year
- [ACL 2025 main] FR-Spec: Frequency-Ranked Speculative Sampling☆36Updated 2 weeks ago
- A flexible and efficient training framework for large-scale alignment tasks☆385Updated last week
- ☆109Updated last month
- [ICLR 2025] COAT: Compressing Optimizer States and Activation for Memory-Efficient FP8 Training☆215Updated 3 weeks ago
- 📰 Must-read papers on KV Cache Compression (constantly updating 🤗).☆481Updated 2 weeks ago