Optimization-AI / fast_clip
☆21Updated 3 months ago
Alternatives and similar repositories for fast_clip:
Users that are interested in fast_clip are comparing it to the libraries listed below
- [ICCV23] Official implementation of eP-ALM: Efficient Perceptual Augmentation of Language Models.☆27Updated last year
- A big_vision inspired repo that implements a generic Auto-Encoder class capable in representation learning and generative modeling.☆34Updated 7 months ago
- [EMNLP 2024] Official code for "Beyond Embeddings: The Promise of Visual Table in Multi-Modal Models"☆14Updated 3 months ago
- Official code for "pi-Tuning: Transferring Multimodal Foundation Models with Optimal Multi-task Interpolation", ICML 2023.☆32Updated last year
- We introduce new approach, Token Reduction using CLIP Metric (TRIM), aimed at improving the efficiency of MLLMs without sacrificing their…☆11Updated last month
- [ACL 2023] PuMer: Pruning and Merging Tokens for Efficient Vision Language Models☆29Updated 3 months ago
- UnifiedMLLM: Enabling Unified Representation for Multi-modal Multi-tasks With Large Language Model☆20Updated 5 months ago
- Code for "Are “Hierarchical” Visual Representations Hierarchical?" in NeurIPS Workshop for Symmetry and Geometry in Neural Representation…☆20Updated last year
- Code for T-MARS data filtering☆35Updated last year
- ☆41Updated 2 weeks ago
- This repository is the implementation of the paper Training Free Pretrained Model Merging (CVPR2024).☆27Updated 10 months ago
- Code for "AVG-LLaVA: A Multimodal Large Model with Adaptive Visual Granularity"☆19Updated 3 months ago
- ☆37Updated 2 months ago
- [ICLR 23] Contrastive Aligned of Vision to Language Through Parameter-Efficient Transfer Learning☆37Updated last year
- ☆15Updated 2 weeks ago
- ☆10Updated 3 months ago
- This repo contains code for "VISTA: Enhancing Long-Duration and High-Resolution Video Understanding by VIdeo SpatioTemporal Augmentation"☆11Updated 3 weeks ago
- ☆36Updated 3 weeks ago
- [NeurIPS-24] This is the official implementation of the paper "DeepStack: Deeply Stacking Visual Tokens is Surprisingly Simple and Effect…☆35Updated 7 months ago
- Code and benchmark for the paper: "A Practitioner's Guide to Continual Multimodal Pretraining" [NeurIPS'24]☆49Updated last month
- Preference Learning for LLaVA☆35Updated 2 months ago
- ☆29Updated last year
- [ECCV 2024] This is the official implementation of "Stitched ViTs are Flexible Vision Backbones".☆25Updated last year
- Original code base for On Pretraining Data Diversity for Self-Supervised Learning☆13Updated last month
- Do Vision and Language Models Share Concepts? A Vector Space Alignment Study☆14Updated 2 months ago
- ☆11Updated 6 months ago
- https://arxiv.org/abs/2209.15162☆49Updated 2 years ago
- ☆31Updated 11 months ago
- Project for SNARE benchmark☆10Updated 7 months ago
- Official Repository of Personalized Visual Instruct Tuning☆26Updated 2 months ago