StarRing2022 / R1-Nature
最简易的R1结果在小模型上的复现,阐述类O1与DeepSeek R1最重要的本质。Think is all your need。利用实验佐证,对于强推理能力,think思考过程性内容是AGI/ASI的核心。
☆45Updated 3 months ago
Alternatives and similar repositories for R1-Nature
Users that are interested in R1-Nature are comparing it to the libraries listed below
Sorting:
- Qwen DianJin: LLMs for the Financial Industry by Alibaba Cloud☆62Updated 2 weeks ago
- Search, organize, discover anything!☆49Updated last year
- ☆94Updated 5 months ago
- ☆36Updated 8 months ago
- Qwen-WisdomVast is a large model trained on 1 million high-quality Chinese multi-turn SFT data, 200,000 English multi-turn SFT data, and …☆18Updated last year
- SkyScript-100M: 1,000,000,000 Pairs of Scripts and Shooting Scripts for Short Drama: https://arxiv.org/abs/2408.09333v2☆119Updated 5 months ago
- SELF-GUIDE: Better Task-Specific Instruction Following via Self-Synthetic Finetuning. COLM 2024 Accepted Paper☆32Updated 11 months ago
- 1.4B sLLM for Chinese and English - HammerLLM🔨☆44Updated last year
- [ICML2025] The official implementation of "C-3PO: Compact Plug-and-Play Proxy Optimization to Achieve Human-like Retrieval-Augmented Gene…☆20Updated last week
- Our 2nd-gen LMM☆33Updated 11 months ago
- Official implementation for "ScoreFlow: Mastering LLM Agent Workflows via Score-based Preference Optimization"☆68Updated 2 months ago
- GLM Series Edge Models☆138Updated 2 months ago
- the newest version of llama3,source code explained line by line using Chinese☆22Updated last year
- Copy the MLP of llama3 8 times as 8 experts , created a router with random initialization,add load balancing loss to construct an 8x8b Mo…☆26Updated 10 months ago
- From Hours to Minutes: Lossless Acceleration of Ultra Long Sequence Generation☆89Updated this week
- A Simple MLLM Surpassed QwenVL-Max with OpenSource Data Only in 14B LLM.☆37Updated 8 months ago
- SUS-Chat: Instruction tuning done right☆48Updated last year
- DeepSolution: Boosting Complex Engineering Solution Design via Tree-based Exploration and Bi-point Thinking☆45Updated 2 months ago
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];☆37Updated last year
- ☆46Updated 10 months ago
- ☆27Updated 3 months ago
- ☆40Updated last year
- Collection of model-centric MCP servers☆14Updated this week
- Knowledge-Reasoning Synergy Reinforcement Learning.☆35Updated 2 months ago
- The complete training code of the open-source high-performance Llama model, including the full process from pre-training to RLHF.☆69Updated 2 years ago
- The source code and dataset mentioned in the paper Seal-Tools: Self-Instruct Tool Learning Dataset for Agent Tuning and Detailed Benchmar…☆48Updated 6 months ago
- 我们是第一个完全可商用的角色大模型。☆40Updated 9 months ago
- ☆47Updated 4 months ago
- ☆29Updated 8 months ago
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆57Updated 5 months ago