ysyisyourbrother / Galaxy-LMLinks
Work in progress LLM framework.
☆15Updated last year
Alternatives and similar repositories for Galaxy-LM
Users that are interested in Galaxy-LM are comparing it to the libraries listed below
Sorting:
- ☆213Updated last year
- ☆102Updated last year
- Survey Paper List - Efficient LLM and Foundation Models☆259Updated last year
- A curated list of awesome projects and papers for AI on Mobile/IoT/Edge devices. Everything is continuously updating. Welcome contributio…☆44Updated 2 years ago
- ☆107Updated last week
- This is a list of awesome edgeAI inference related papers.☆98Updated 2 years ago
- ☆172Updated last year
- InFi is a library for building input filters for resource-efficient inference.☆41Updated 2 years ago
- Curated collection of papers in MoE model inference☆329Updated 2 months ago
- ☆16Updated 2 years ago
- This repository serves as a comprehensive survey of LLM development, featuring numerous research papers along with their corresponding co…☆264Updated last month
- Split Learning Simulation Framework for LLMs☆36Updated last year
- One-size-fits-all model for mobile AI, a novel paradigm for mobile AI in which the OS and hardware co-manage a foundation model that is c…☆29Updated last year
- This repository is established to store personal notes and annotated papers during daily research.☆173Updated this week
- THC: Accelerating Distributed Deep Learning Using Tensor Homomorphic Compression☆20Updated last year
- [ASPLOS'25] Towards End-to-End Optimization of LLM-based Applications with Ayo☆57Updated 5 months ago
- Source code for Jellyfish, a soft real-time inference serving system☆14Updated 3 years ago
- zTT: Learning-based DVFS with Zero Thermal Throttling for Mobile Devices [MobiSys'21] - Artifact Evaluation☆28Updated 4 years ago
- Here are my personal paper reading notes (including cloud computing, resource management, systems, machine learning, deep learning, and o…☆146Updated this week
- ☆22Updated last year
- Summary of some awesome work for optimizing LLM inference☆162Updated last month
- ☆47Updated last year
- ☆319Updated last year
- Efficient Interactive LLM Serving with Proxy Model-based Sequence Length Prediction | A tiny BERT model can tell you the verbosity of an …☆48Updated last year
- ☆15Updated 2 years ago
- Enabling High Quality Real-Time Communications with Adaptive Frame-Rate (USENIX NSDI 2023)☆23Updated 2 years ago
- a curated list of high-quality papers on resource-efficient LLMs 🌱☆154Updated 9 months ago
- Awesome-LLM-KV-Cache: A curated list of 📙Awesome LLM KV Cache Papers with Codes.☆405Updated 10 months ago
- ☆15Updated 11 months ago
- Accommodating Large Language Model Training over Heterogeneous Environment.☆25Updated 9 months ago