dinobby / Symbolic-MoELinks
The code implementation of Symbolic-MoE
☆36Updated 5 months ago
Alternatives and similar repositories for Symbolic-MoE
Users that are interested in Symbolic-MoE are comparing it to the libraries listed below
Sorting:
- JudgeLRM: Large Reasoning Models as a Judge☆32Updated 3 months ago
- OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement☆104Updated 2 weeks ago
- ☆47Updated 5 months ago
- Official implementation of the paper "Soft Thinking: Unlocking the Reasoning Potential of LLMs in Continuous Concept Space"☆204Updated 2 weeks ago
- [ACL 2025] A Generalizable and Purely Unsupervised Self-Training Framework☆68Updated 2 months ago
- ☆212Updated 5 months ago
- ☆61Updated 5 months ago
- [ICCV 2025] Auto Interpretation Pipeline and many other functionalities for Multimodal SAE Analysis.☆146Updated 3 weeks ago
- codes for R-Zero: Self-Evolving Reasoning LLM from Zero Data (https://www.arxiv.org/pdf/2508.05004)☆128Updated this week
- ☆126Updated 2 months ago
- Code for "Your Mixture-of-Experts LLM Is Secretly an Embedding Model For Free"☆76Updated 9 months ago
- Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"☆120Updated 9 months ago
- 📖 This is a repository for organizing papers, codes, and other resources related to Latent Reasoning.☆176Updated this week
- [NeurIPS 2024] A task generation and model evaluation system for multimodal language models.☆72Updated 8 months ago
- ☆34Updated 2 months ago
- Process Reward Models That Think☆47Updated last month
- SIFT: Grounding LLM Reasoning in Contexts via Stickers☆57Updated 5 months ago
- Code for Heima☆51Updated 3 months ago
- [ACL 2025] An inference-time decoding strategy with adaptive foresight sampling☆101Updated 2 months ago
- [ACL 2025] Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆99Updated last month
- official code for "BoostStep: Boosting mathematical capability of Large Language Models via improved single-step reasoning"☆36Updated 6 months ago
- ☆128Updated 3 months ago
- Large Language Models Can Self-Improve in Long-context Reasoning☆72Updated 8 months ago
- [ICLR 2025] SuperCorrect: Advancing Small LLM Reasoning with Thought Template Distillation and Self-Correction☆76Updated 4 months ago
- Official repository for paper: O1-Pruner: Length-Harmonizing Fine-Tuning for O1-Like Reasoning Pruning☆86Updated 5 months ago
- Repo for "Z1: Efficient Test-time Scaling with Code"☆63Updated 3 months ago
- Official code repository for Sketch-of-Thought (SoT)☆125Updated 3 months ago
- Chain of Experts (CoE) enables communication between experts within Mixture-of-Experts (MoE) models☆219Updated last month
- Official Implementation of ARPO: End-to-End Policy Optimization for GUI Agents with Experience Replay☆104Updated 2 months ago
- ☆81Updated 2 weeks ago