bytedance / SandboxFusion
☆160Updated last month
Alternatives and similar repositories for SandboxFusion:
Users that are interested in SandboxFusion are comparing it to the libraries listed below
- A lightweight reproduction of DeepSeek-R1-Zero with indepth analysis of self-reflection behavior.☆212Updated this week
- A Comprehensive Survey on Long Context Language Modeling☆86Updated last week
- A Comprehensive Benchmark for Software Development.☆100Updated 9 months ago
- A flexible and efficient training framework for large-scale alignment tasks☆333Updated last month
- Codes for the paper "∞Bench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718☆313Updated 6 months ago
- Reproducing R1 for Code with Reliable Rewards☆132Updated 3 weeks ago
- Offical Repo for "Programming Every Example: Lifting Pre-training Data Quality Like Experts at Scale"☆229Updated last month
- A visuailzation tool to make deep understaning and easier debugging for RLHF training.☆177Updated last month
- ☆318Updated 8 months ago
- CodeRAG-Bench: Can Retrieval Augment Code Generation?☆119Updated 4 months ago
- ☆60Updated 4 months ago
- [ACL 2024 Demo] Official GitHub repo for UltraEval: An open source framework for evaluating foundation models.☆237Updated 4 months ago
- The related works and background techniques about Openai o1☆217Updated 2 months ago
- Codev-Bench (Code Development Benchmark), a fine-grained, real-world, repository-level, and developer-centric evaluation framework. Codev…☆38Updated 4 months ago
- ACL 2024 | LooGLE: Long Context Evaluation for Long-Context Language Models☆179Updated 5 months ago
- A new tool learning benchmark aiming at well-balanced stability and reality, based on ToolBench.☆135Updated 3 weeks ago
- The repository for paper "DebugBench: "Evaluating Debugging Capability of Large Language Models".☆68Updated 8 months ago
- Official repository for our paper "FullStack Bench: Evaluating LLMs as Full Stack Coders"☆72Updated 3 months ago
- ☆312Updated 6 months ago
- Inference code of Lingma SWE-GPT☆199Updated 3 months ago
- ☆263Updated 8 months ago
- CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion (NeurIPS 2023)☆136Updated 7 months ago
- A series of technical report on Slow Thinking with LLM☆581Updated this week
- Exploring the Limit of Outcome Reward for Learning Mathematical Reasoning☆158Updated last week
- ☆124Updated 3 weeks ago
- GPT-Fathom is an open-source and reproducible LLM evaluation suite, benchmarking 10+ leading open-source and closed-source LLMs as well a…☆349Updated 11 months ago
- R1-searcher: Incentivizing the Search Capability in LLMs via Reinforcement Learning☆376Updated this week
- Super-Efficient RLHF Training of LLMs with Parameter Reallocation☆242Updated 2 months ago
- A simple toolkit for benchmarking LLMs on mathematical reasoning tasks. 🧮✨☆191Updated 11 months ago
- Repo of paper "Free Process Rewards without Process Labels"☆138Updated last week