nex-agi / NexGAPLinks
Nex General Agentic Data Pipeline, an end-to-end pipeline for generating high-quality agentic training data.
☆28Updated last month
Alternatives and similar repositories for NexGAP
Users that are interested in NexGAP are comparing it to the libraries listed below
Sorting:
- Super-Efficient RLHF Training of LLMs with Parameter Reallocation☆330Updated 8 months ago
- ☆97Updated last month
- PoC for "SpecReason: Fast and Accurate Inference-Time Compute via Speculative Reasoning" [NeurIPS '25]☆61Updated 3 months ago
- Implementation for FP8/INT8 Rollout for RL training without performence drop.☆281Updated 2 months ago
- Bridge Megatron-Core to Hugging Face/Reinforcement Learning☆181Updated this week
- ByteCheckpoint: An Unified Checkpointing Library for LFMs☆258Updated last month
- Official Implementation of "Learning Harmonized Representations for Speculative Sampling" (HASS)☆52Updated 9 months ago
- End-to-End Reinforcement Learning for Multi-Turn Tool-Integrated Reasoning☆348Updated 3 months ago
- NexRL is an ultra-loosely-coupled LLM post-training framework.☆63Updated last month
- Spec-Bench: A Comprehensive Benchmark and Unified Evaluation Platform for Speculative Decoding (ACL 2024 Findings)☆350Updated 8 months ago
- [ICLR 2025] PEARL: Parallel Speculative Decoding with Adaptive Draft Length☆144Updated 2 weeks ago
- ☆64Updated last year
- Evaluation utilities based on SymPy.☆21Updated last year
- [ASPLOS'26] Taming the Long-Tail: Efficient Reasoning RL Training with Adaptive Drafter☆121Updated last month
- 使用torch.distributed实现DP/TP/PP☆12Updated 2 years ago
- ☆208Updated 2 months ago
- ☆49Updated last year
- (best/better) practices of megatron on veRL and tuning guide☆114Updated 3 months ago
- [ICLR 2025🔥] D2O: Dynamic Discriminative Operations for Efficient Long-Context Inference of Large Language Models☆25Updated 6 months ago
- Chain of Thoughts (CoT) is so hot! so long! We need short reasoning process!☆71Updated 9 months ago
- Code associated with the paper **Draft & Verify: Lossless Large Language Model Acceleration via Self-Speculative Decoding**☆214Updated 10 months ago
- [NeurIPS'23] H2O: Heavy-Hitter Oracle for Efficient Generative Inference of Large Language Models.☆498Updated last year
- Codes for the paper "BAPO: Stabilizing Off-Policy Reinforcement Learning for LLMs via Balanced Policy Optimization with Adaptive Clipping…☆89Updated 2 months ago
- Official PyTorch implementation of the paper "dLLM-Cache: Accelerating Diffusion Large Language Models with Adaptive Caching" (dLLM-Cache…☆191Updated last month
- [ACL 2025 main] FR-Spec: Frequency-Ranked Speculative Sampling☆49Updated 5 months ago
- ☆90Updated 6 months ago
- [ICML 2024] Quest: Query-Aware Sparsity for Efficient Long-Context LLM Inference☆362Updated 6 months ago
- ☆443Updated 5 months ago
- Train speculative decoding models effortlessly and port them smoothly to SGLang serving.☆615Updated this week
- Reasoning or Memorization? Unreliable Results of Reinforcement Learning Due to Data Contamination.☆21Updated 5 months ago