inclusionAI / RingLinks
Ring is a reasoning MoE LLM provided and open-sourced by InclusionAI, derived from Ling.
☆81Updated last week
Alternatives and similar repositories for Ring
Users that are interested in Ring are comparing it to the libraries listed below
Sorting:
- ☆50Updated last week
- ☆85Updated 2 weeks ago
- ☆107Updated 3 weeks ago
- ☆51Updated 3 months ago
- ☆87Updated 2 months ago
- ☆47Updated 2 weeks ago
- This is the official repo of "QuickLLaMA: Query-aware Inference Acceleration for Large Language Models"☆53Updated 11 months ago
- ☆116Updated last month
- The official implementation of paper: SimLayerKV: A Simple Framework for Layer-Level KV Cache Reduction.☆46Updated 8 months ago
- Repo for "Z1: Efficient Test-time Scaling with Code"☆61Updated 2 months ago
- ☆55Updated last week
- ☆58Updated last week
- A lightweight reinforcement learning framework that integrates seamlessly into your codebase, empowering developers to focus on algorithm…☆32Updated last month
- [ICML 2025] Predictive Data Selection: The Data That Predicts Is the Data That Teaches☆50Updated 3 months ago
- "what, how, where, and how well? a survey on test-time scaling in large language models" repository☆45Updated this week
- The official repo of SynLogic: Synthesizing Verifiable Reasoning Data at Scale for Learning Logical Reasoning and Beyond☆147Updated 3 weeks ago
- A Sober Look at Language Model Reasoning☆74Updated last week
- ☆24Updated 3 months ago
- [ICLR 2025] MiniPLM: Knowledge Distillation for Pre-Training Language Models☆47Updated 7 months ago
- qwen-nsa☆67Updated 2 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆85Updated 9 months ago
- Long Context Extension and Generalization in LLMs☆57Updated 9 months ago
- ARM: Adaptive Reasoning Model☆43Updated last week
- Code for ICLR 2025 Paper "What is Wrong with Perplexity for Long-context Language Modeling?"☆90Updated last month
- Trinity-RFT is a general-purpose, flexible and scalable framework designed for reinforcement fine-tuning (RFT) of large language models (…☆129Updated this week
- RL Scaling and Test-Time Scaling (ICML'25)☆107Updated 5 months ago
- A comrephensive collection of learning from rewards in the post-training and test-time scaling of LLMs, with a focus on both reward model…☆47Updated 2 weeks ago
- The official implementation for Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free☆44Updated last month
- The source code of "Merging Experts into One: Improving Computational Efficiency of Mixture of Experts (EMNLP 2023)":☆39Updated last year
- Revisiting Mid-training in the Era of RL Scaling☆62Updated 2 months ago