mlfoundations / MINT-1T
MINT-1T: A one trillion token multimodal interleaved dataset.
☆797Updated 6 months ago
Alternatives and similar repositories for MINT-1T:
Users that are interested in MINT-1T are comparing it to the libraries listed below
- From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)☆633Updated 3 months ago
- ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Expert…☆1,349Updated 2 months ago
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation☆718Updated 6 months ago
- ☆598Updated last year
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.☆698Updated 4 months ago
- [ICML 2024] CLLMs: Consistency Large Language Models☆371Updated 3 months ago
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Models☆1,445Updated 11 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆590Updated last month
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.☆916Updated 8 months ago
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,359Updated 10 months ago
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills☆722Updated last year
- [NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333☆1,081Updated last year
- PyTorch Implementation of "V* : Guided Visual Search as a Core Mechanism in Multimodal LLMs"☆565Updated last year
- LLaVA-UHD v2: an MLLM Integrating High-Resolution Feature Pyramid via Hierarchical Window Transformer☆364Updated last month
- Minimalistic large language model 3D-parallelism training☆1,445Updated this week
- Reference implementation of Megalodon 7B model☆514Updated 9 months ago
- Recipes to scale inference-time compute of open models☆992Updated last month
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning☆638Updated 8 months ago
- Official implementation of "Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling"☆840Updated 3 weeks ago
- Code for Quiet-STaR☆711Updated 5 months ago
- ☆496Updated 2 months ago
- [NeurIPS'24 Spotlight, ICLR'25] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which r…☆907Updated this week
- LLM2CLIP makes SOTA pretrained CLIP model more SOTA ever.☆465Updated 3 weeks ago
- Large Reasoning Models☆802Updated 2 months ago
- Yi-1.5 is an upgraded version of Yi, delivering stronger performance in coding, math, reasoning, and instruction-following capability.☆548Updated 3 months ago
- TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones☆1,260Updated 9 months ago
- A Framework of Small-scale Large Multimodal Models☆740Updated 2 weeks ago
- A family of lightweight multimodal models.☆986Updated 2 months ago
- Reaching LLaMA2 Performance with 0.1M Dollars☆972Updated 6 months ago