dinobby / Symbolic-MoE
The code implementation of Symbolic-MoE
☆27Updated last month
Alternatives and similar repositories for Symbolic-MoE:
Users that are interested in Symbolic-MoE are comparing it to the libraries listed below
- ☆24Updated last week
- Resources for our paper: "EvoAgent: Towards Automatic Multi-Agent Generation via Evolutionary Algorithms"☆93Updated 6 months ago
- ☆22Updated last week
- official code for "BoostStep: Boosting mathematical capability of Large Language Models via improved single-step reasoning"☆35Updated 3 months ago
- ☆13Updated 4 months ago
- Code for "Your Mixture-of-Experts LLM Is Secretly an Embedding Model For Free"☆61Updated 6 months ago
- Code for Heima☆40Updated this week
- ☆86Updated 2 weeks ago
- ☆55Updated 2 months ago
- ☆24Updated last month
- OpenVLThinker: An Early Exploration to Vision-Language Reasoning via Iterative Self-Improvement☆71Updated 3 weeks ago
- Official Repository of Are Your LLMs Capable of Stable Reasoning?☆25Updated last month
- ☆15Updated 2 weeks ago
- Repo for "Z1: Efficient Test-time Scaling with Code"☆55Updated 2 weeks ago
- ☆46Updated 2 months ago
- [EMNLP 2024 Findings] ProSA: Assessing and Understanding the Prompt Sensitivity of LLMs☆24Updated 6 months ago
- Agentic Knowledgeable Self-awareness☆50Updated last week
- ☆36Updated last month
- ☆63Updated this week
- ☆24Updated 7 months ago
- The official implementation of Self-Exploring Language Models (SELM)☆63Updated 10 months ago
- Official repo of paper LM2☆37Updated 2 months ago
- Code Implementation, Evaluations, Documentation, Links and Resources for Min P paper☆32Updated last month
- ☆31Updated 3 months ago
- Agentic Reward Modeling: Integrating Human Preferences with Verifiable Correctness Signals for Reliable Reward Systems☆86Updated last month
- Unofficial Implementation of Chain-of-Thought Reasoning Without Prompting☆32Updated last year
- The official repository for SkyLadder: Better and Faster Pretraining via Context Window Scheduling☆29Updated last month
- [Preprint] A Generalizable and Purely Unsupervised Self-Training Framework☆50Updated last week
- What Happened in LLMs Layers when Trained for Fast vs. Slow Thinking: A Gradient Perspective☆63Updated last month
- [ICLR 2025] Video-STaR: Self-Training Enables Video Instruction Tuning with Any Supervision☆60Updated 9 months ago