deepseek-ai / EngramLinks
Conditional Memory via Scalable Lookup: A New Axis of Sparsity for Large Language Models
β1,612Updated this week
Alternatives and similar repositories for Engram
Users that are interested in Engram are comparing it to the libraries listed below
Sorting:
- β816Updated 7 months ago
- π³ Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"β952Updated 9 months ago
- Muon is Scalable for LLM Trainingβ1,397Updated 5 months ago
- Parallel Scaling Law for Language Model β Beyond Parameter and Inference Time Scalingβ467Updated 7 months ago
- β443Updated 5 months ago
- β1,421Updated last month
- slime is an LLM post-training framework for RL Scaling.β3,224Updated last week
- MoBA: Mixture of Block Attention for Long-Context LLMsβ2,031Updated 9 months ago
- VeOmni: Scaling Any Modality Model Training with Model-Centric Distributed Recipe Zooβ1,524Updated this week
- β1,260Updated last month
- β1,257Updated last month
- Checkpoint-engine is a simple middleware to update model weights in LLM inference enginesβ888Updated this week
- A construction kit for reinforcement learning environment management.β292Updated this week
- β1,515Updated last month
- DeepConf: Deep Think with Confidenceβ352Updated 3 months ago
- Speed Always Wins: A Survey on Efficient Architectures for Large Language Modelsβ385Updated 2 months ago
- Seed-Coder is a family of lightweight open-source code LLMs comprising base, instruct and reasoning models, developed by ByteDance Seed.β725Updated 7 months ago
- A MemAgent framework that can be extrapolated to 3.5M, along with a training framework for RL training of any agent workflow.β849Updated 5 months ago
- Ling is a MoE LLM provided and open-sourced by InclusionAI.β238Updated 8 months ago
- A framework for efficient model inference with omni-modality modelsβ2,087Updated this week
- β1,376Updated 4 months ago
- β208Updated 2 months ago
- An Efficient and User-Friendly Scaling Library for Reinforcement Learning with Large Language Modelsβ2,619Updated this week
- Unleashing the Power of Reinforcement Learning for Math and Code Reasonersβ737Updated 7 months ago
- Official implementation of "Fast-dLLM: Training-free Acceleration of Diffusion LLM by Enabling KV Cache and Parallel Decoding"β773Updated last month
- Miles is an enterprise-facing reinforcement learning framework for large-scale MoE post-training and production workloads, forked from anβ¦β714Updated this week
- Lightning-Fast RL for LLM Reasoning and Agents. Made Simple & Flexible.β3,381Updated this week
- β508Updated 3 weeks ago
- dInfer: An Efficient Inference Framework for Diffusion Language Modelsβ389Updated last week
- β870Updated last month