mlfoundations / MINT-1T
MINT-1T: A one trillion token multimodal interleaved dataset.
☆808Updated 8 months ago
Alternatives and similar repositories for MINT-1T:
Users that are interested in MINT-1T are comparing it to the libraries listed below
- Understanding R1-Zero-Like Training: A Critical Perspective☆882Updated last week
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation☆751Updated 8 months ago
- From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)☆692Updated 5 months ago
- [ICML 2024] CLLMs: Consistency Large Language Models☆390Updated 5 months ago
- LLM Analytics☆655Updated 6 months ago
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills☆739Updated last year
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,378Updated last year
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.☆1,288Updated this week
- Yi-1.5 is an upgraded version of Yi, delivering stronger performance in coding, math, reasoning, and instruction-following capability.☆553Updated 5 months ago
- TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones☆1,280Updated last year
- OLMoE: Open Mixture-of-Experts Language Models☆723Updated last month
- DataComp for Language Models☆1,279Updated last month
- Unleashing the Power of Reinforcement Learning for Math and Code Reasoners☆517Updated this week
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.☆719Updated 7 months ago
- ☆610Updated last year
- Recipes to scale inference-time compute of open models☆1,058Updated 2 months ago
- Reaching LLaMA2 Performance with 0.1M Dollars☆981Updated 9 months ago
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.☆927Updated last month
- LLaVA-UHD v2: an MLLM Integrating High-Resolution Semantic Pyramid via Hierarchical Window Transformer☆374Updated last week
- Repository for Meta Chameleon, a mixed-modal early-fusion foundation model from FAIR.☆1,990Updated 8 months ago
- Muon is Scalable for LLM Training☆1,029Updated 3 weeks ago
- Large Reasoning Models☆802Updated 4 months ago
- LIMO: Less is More for Reasoning☆920Updated 3 weeks ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆865Updated 2 months ago
- [NeurIPS'24 Spotlight, ICLR'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which r…☆985Updated this week
- Codebase for Merging Language Models (ICML 2024)☆818Updated 11 months ago
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Models☆1,514Updated last year
- Next-Token Prediction is All You Need☆2,099Updated last month
- Serving multiple LoRA finetuned LLM as one☆1,054Updated 11 months ago
- This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for E…☆420Updated last week