mlfoundations / MINT-1TLinks
π MINT-1T: A one trillion token multimodal interleaved dataset.
β828Updated last year
Alternatives and similar repositories for MINT-1T
Users that are interested in MINT-1T are comparing it to the libraries listed below
Sorting:
- Code release for "LLMs can see and hear without any training"β458Updated 8 months ago
- Reference implementation of Megalodon 7B modelβ529Updated 8 months ago
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.β951Updated 10 months ago
- TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbonesβ1,306Updated last year
- PyTorch Implementation of "V* : Guided Visual Search as a Core Mechanism in Multimodal LLMs"β685Updated 2 years ago
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.β1,406Updated 9 months ago
- NeurIPS 2025 Spotlight; ICLR2024 Spotlight; CVPR 2024; EMNLP 2024β1,807Updated 2 months ago
- From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)β789Updated last year
- Yi-1.5 is an upgraded version of Yi, delivering stronger performance in coding, math, reasoning, and instruction-following capability.β557Updated last year
- Understanding R1-Zero-Like Training: A Critical Perspectiveβ1,203Updated 5 months ago
- LLM Analyticsβ705Updated last year
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAIβ1,407Updated last year
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generationβ823Updated 7 months ago
- OLMoE: Open Mixture-of-Experts Language Modelsβ961Updated 4 months ago
- [COLM 2025] LIMO: Less is More for Reasoningβ1,061Updated 6 months ago
- [ICML 2024] CLLMs: Consistency Large Language Modelsβ411Updated last year
- β642Updated last year
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skillsβ763Updated 2 years ago
- Unleashing the Power of Reinforcement Learning for Math and Code Reasonersβ741Updated 7 months ago
- LLaVA-UHD v3: Progressive Visual Compression for Efficient Native-Resolution Encoding in MLLMsβ413Updated last month
- HPT - Open Multimodal LLMs from HyperGAIβ315Updated last year
- Inference code for Persimmon-8Bβ412Updated 2 years ago
- DataDreamer: Prompt. Generate Synthetic Data. Train & Align Models. β π€π€β1,088Updated last year
- Open weights language model from Google DeepMind, based on Griffin.β661Updated last week
- This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for Eβ¦β544Updated 8 months ago
- Reaching LLaMA2 Performance with 0.1M Dollarsβ987Updated last year
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Modelsβ1,655Updated last year
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β750Updated last year
- [ICLR2025 Spotlightπ₯] Official Implementation of TokenFormer: Rethinking Transformer Scaling with Tokenized Model Parametersβ584Updated 11 months ago
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuningβ668Updated last year