gaasher / I-JEPALinks
Implementation of I-JEPA from "Self-Supervised Learning from Images with a Joint-Embedding Predictive Architecture"
☆281Updated last year
Alternatives and similar repositories for I-JEPA
Users that are interested in I-JEPA are comparing it to the libraries listed below
Sorting:
- Fine-tuning "ImageBind One Embedding Space to Bind Them All" with LoRA☆193Updated 2 years ago
- Official code for "TOAST: Transfer Learning via Attention Steering"☆188Updated 2 years ago
- Implementation of Soft MoE, proposed by Brain's Vision team, in Pytorch☆340Updated 9 months ago
- Implementation of Recurrent Memory Transformer, Neurips 2022 paper, in Pytorch☆422Updated last year
- Implementation of Block Recurrent Transformer - Pytorch☆223Updated last year
- Experiments in Joint Embedding Predictive Architectures (JEPAs).☆45Updated 2 years ago
- Code release for "Dropout Reduces Underfitting"☆317Updated 2 years ago
- Official JAX implementation of Learning to (Learn at Test Time): RNNs with Expressive Hidden States☆445Updated 2 months ago
- PyTorch Implementation of Jamba: "Jamba: A Hybrid Transformer-Mamba Language Model"☆203Updated this week
- Internet Explorer explores the web in a self-supervised manner to progressively find relevant examples that improve performance on a desi…☆163Updated 2 years ago
- A framework for merging models solving different tasks with different initializations into one multi-task model without any additional tr…☆309Updated last year
- This is the official repository for the LENS (Large Language Models Enhanced to See) system.☆356Updated 5 months ago
- Trying out the Mamba architecture on small examples (cifar-10, shakespeare char level etc.)☆47Updated 2 years ago
- Hiera: A fast, powerful, and simple hierarchical vision transformer.☆1,048Updated last year
- ☆210Updated 2 years ago
- Learning from synthetic data - code and models☆326Updated 2 years ago
- Experiments around a simple idea for inducing multiple hierarchical predictive model within a GPT☆224Updated last year
- Implementation of 🌻 Mirasol, SOTA Multimodal Autoregressive model out of Google Deepmind, in Pytorch☆91Updated 2 years ago
- ☆190Updated 2 years ago
- Code release for "Learning Video Representations from Large Language Models"☆537Updated 2 years ago
- Code repository for Black Mamba☆260Updated last year
- When do we not need larger vision models?☆413Updated 11 months ago
- A novel implementation of fusing ViT with Mamba into a fast, agile, and high performance Multi-Modal Model. Powered by Zeta, the simplest…☆462Updated 2 months ago
- Visualizing representations with diffusion based conditional generative model.☆103Updated 2 years ago
- Holds code for our CVPR'23 tutorial: All Things ViTs: Understanding and Interpreting Attention in Vision.☆196Updated 2 years ago
- Official PyTorch Implementation of "The Hidden Attention of Mamba Models"☆232Updated 2 months ago
- [TMLR23] Official implementation of UnIVAL: Unified Model for Image, Video, Audio and Language Tasks.☆232Updated 2 years ago
- Projects based on SigLIP (Zhai et. al, 2023) and Hugging Face transformers integration 🤗☆297Updated 10 months ago
- Implementation of ST-Moe, the latest incarnation of MoE after years of research at Brain, in Pytorch☆375Updated last year
- Implementation of Lumiere, SOTA text-to-video generation from Google Deepmind, in Pytorch☆281Updated last year