The official repo of MiniMax-Text-01 and MiniMax-VL-01, large-language-model & vision-language-model based on Linear Attention
☆3,362Jul 7, 2025Updated 8 months ago
Alternatives and similar repositories for MiniMax-01
Users that are interested in MiniMax-01 are comparing it to the libraries listed below
Sorting:
- ☆3,472Mar 7, 2025Updated last year
- MoBA: Mixture of Block Attention for Long-Context LLMs☆2,076Apr 3, 2025Updated 11 months ago
- 🚀 Efficient implementations of state-of-the-art linear attention models☆4,630Updated this week
- MiniMax-M1, the world's first open-weight, large-scale hybrid-attention reasoning model.☆3,115Jul 7, 2025Updated 8 months ago
- ☆810Jun 9, 2025Updated 9 months ago
- Explore these applications integrating MiniMax's multimodal API to see how text, vision, and speech processing capabilities are incorpora…☆68Jan 30, 2026Updated last month
- FlashMLA: Efficient Multi-head Latent Attention Kernels☆12,512Feb 6, 2026Updated last month
- Qwen3 is the large language model series developed by Qwen team, Alibaba Cloud.☆26,899Jan 9, 2026Updated 2 months ago
- Muon is Scalable for LLM Training☆1,444Aug 3, 2025Updated 7 months ago
- SGLang is a high-performance serving framework for large language models and multimodal models.☆24,455Updated this week
- ☆3,177Mar 17, 2025Updated last year
- Qwen3-VL is the multimodal large language model series developed by Qwen team, Alibaba Cloud.☆18,671Jan 30, 2026Updated last month
- Fully open reproduction of DeepSeek-R1☆25,941Nov 24, 2025Updated 3 months ago
- Sky-T1: Train your own O1 preview model within $450☆3,369Jul 12, 2025Updated 8 months ago
- A bidirectional pipeline parallelism algorithm for computation-communication overlap in DeepSeek V3/R1 training.☆2,934Jan 14, 2026Updated 2 months ago
- Next-Token Prediction is All You Need☆2,370Jan 12, 2026Updated 2 months ago
- Qwen2.5-Omni is an end-to-end multimodal model by Qwen team at Alibaba Cloud, capable of understanding text, audio, vision, video, and pe…☆3,950Jun 12, 2025Updated 9 months ago
- verl: Volcano Engine Reinforcement Learning for LLMs☆19,919Updated this week
- Kimi-VL: Mixture-of-Experts Vision-Language Model for Multimodal Reasoning, Long-Context Understanding, and Strong Agent Capabilities☆1,164Jul 15, 2025Updated 8 months ago
- Janus-Series: Unified Multimodal Understanding and Generation Models☆17,707Feb 1, 2025Updated last year
- Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation☆7,974May 15, 2025Updated 10 months ago
- Official Repo for Open-Reasoner-Zero☆2,085Jun 2, 2025Updated 9 months ago
- DeepEP: an efficient expert-parallel communication library☆9,044Feb 9, 2026Updated last month
- A Gemini 2.5 Flash Level MLLM for Vision, Speech, and Full-Duplex Multimodal Live Streaming on Your Phone☆24,144Mar 7, 2026Updated last week
- Fast and memory-efficient exact attention☆22,832Updated this week
- An Easy-to-use, Scalable and High-performance Agentic RL Framework based on Ray (PPO & DAPO & REINFORCE++ & TIS & vLLM & Ray & Async RL)☆9,191Updated this week
- Simple RL training for reasoning☆3,834Dec 23, 2025Updated 2 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆73,479Updated this week
- Seed1.5-VL, a vision-language foundation model designed to advance general-purpose multimodal understanding and reasoning, achieving stat…☆1,555Jun 14, 2025Updated 9 months ago
- An Open Large Reasoning Model for Real-World Solutions☆1,539Feb 13, 2026Updated last month
- s1: Simple test-time scaling☆6,642Jun 25, 2025Updated 8 months ago
- DeepGEMM: clean and efficient FP8 GEMM kernels with fine-grained scaling☆6,253Feb 27, 2026Updated 2 weeks ago
- Lightning Attention-2: A Free Lunch for Handling Unlimited Sequence Lengths in Large Language Models☆341Feb 23, 2025Updated last year
- DeepSeek-V2: A Strong, Economical, and Efficient Mixture-of-Experts Language Model☆4,998Sep 25, 2024Updated last year
- Democratizing Reinforcement Learning for LLMs☆5,219Updated this week
- 🐳 Efficient Triton implementations for "Native Sparse Attention: Hardware-Aligned and Natively Trainable Sparse Attention"☆975Feb 5, 2026Updated last month
- Minimal reproduction of DeepSeek R1-Zero☆12,932Feb 27, 2026Updated 2 weeks ago
- Ongoing research training transformer models at scale☆15,647Updated this week
- Agent framework and applications built upon Qwen>=3.0, featuring Function Calling, MCP, Code Interpreter, RAG, Chrome extension, etc.☆15,597Mar 4, 2026Updated 2 weeks ago