mlfoundations / MINT-1TLinks
MINT-1T: A one trillion token multimodal interleaved dataset.
☆816Updated 10 months ago
Alternatives and similar repositories for MINT-1T
Users that are interested in MINT-1T are comparing it to the libraries listed below
Sorting:
- From scratch implementation of a sparse mixture of experts language model inspired by Andrej Karpathy's makemore :)☆715Updated 7 months ago
- Understanding R1-Zero-Like Training: A Critical Perspective☆973Updated 2 weeks ago
- MobileLLM Optimizing Sub-billion Parameter Language Models for On-Device Use Cases. In ICML 2024.☆1,301Updated last month
- Anole: An Open, Autoregressive and Native Multimodal Models for Interleaved Image-Text Generation☆767Updated 10 months ago
- Implementation of the training framework proposed in Self-Rewarding Language Model, from MetaAI☆1,387Updated last year
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.☆729Updated 8 months ago
- [ICML 2024] CLLMs: Consistency Large Language Models☆392Updated 6 months ago
- A family of open-sourced Mixture-of-Experts (MoE) Large Language Models☆1,538Updated last year
- Large Reasoning Models☆803Updated 6 months ago
- Reference implementation of Megalodon 7B model☆520Updated 3 weeks ago
- LLaVA-Plus: Large Language and Vision Assistants that Plug and Learn to Use Skills☆744Updated last year
- MultimodalC4 is a multimodal extension of c4 that interleaves millions of images with text.☆931Updated 2 months ago
- [ICLR 2025] Samba: Simple Hybrid State Space Models for Efficient Unlimited Context Language Modeling☆879Updated last month
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuning☆652Updated last year
- TinyGPT-V: Efficient Multimodal Large Language Model via Small Backbones☆1,287Updated last year
- YaRN: Efficient Context Window Extension of Large Language Models☆1,495Updated last year
- LIMO: Less is More for Reasoning☆955Updated 2 months ago
- LLM Analytics☆664Updated 7 months ago
- Serving multiple LoRA finetuned LLM as one☆1,062Updated last year
- LLaVA-UHD v2: an MLLM Integrating High-Resolution Semantic Pyramid via Hierarchical Window Transformer☆379Updated last month
- This repo contains evaluation code for the paper "MMMU: A Massive Multi-discipline Multimodal Understanding and Reasoning Benchmark for E…☆438Updated 3 weeks ago
- Recipes to scale inference-time compute of open models☆1,090Updated 2 weeks ago
- Code release for "LLMs can see and hear without any training"☆438Updated last month
- Code for Quiet-STaR☆732Updated 9 months ago
- ICLR2024 Spotlight: curation/training code, metadata, distribution and pre-trained models for MetaCLIP; CVPR 2024: MoDE: CLIP Data Expert…☆1,450Updated 2 months ago
- HPT - Open Multimodal LLMs from HyperGAI☆316Updated last year
- Reaching LLaMA2 Performance with 0.1M Dollars☆981Updated 10 months ago
- OLMoE: Open Mixture-of-Experts Language Models☆773Updated 2 months ago
- [NeurIPS 2023] MeZO: Fine-Tuning Language Models with Just Forward Passes. https://arxiv.org/abs/2305.17333☆1,113Updated last year
- A family of lightweight multimodal models.☆1,020Updated 6 months ago