DAILtech / LLaVA-Deploy-GuideLinks
💻 Tutorial for deploying LLaVA (Large Language & Vision Assistant) on Ubuntu + CUDA – step-by-step guide with CLI & web UI.
☆18Updated 9 months ago
Alternatives and similar repositories for LLaVA-Deploy-Guide
Users that are interested in LLaVA-Deploy-Guide are comparing it to the libraries listed below
Sorting:
- ☆19Updated last year
- ☆43Updated 10 months ago
- [AAAI 2024] Prompt-based Distribution Alignment for Unsupervised Domain Adaptation☆78Updated last year
- The official pytorch implemention of our CVPR-2024 paper "MMA: Multi-Modal Adapter for Vision-Language Models".☆95Updated 9 months ago
- [CVPR 2025] Official implementation of paper "Multi-Granularity Class Prototype Topology Distillation for Class-Incremental Source-Free …☆17Updated 5 months ago
- Easy wrapper for inserting LoRA layers in CLIP.☆40Updated last year
- The repo for "Enhancing Multi-modal Cooperation via Sample-level Modality Valuation", CVPR 2024☆59Updated last year
- Deep Correlated Prompting for Visual Recognition with Missing Modalities (NeurIPS 2024)☆34Updated 11 months ago
- PMR: Prototypical Modal Rebalance for Multimodal Learning☆45Updated 2 years ago
- [EMNLP 2024] Implementation of vision-language model fine-tuning via simple parameter-efficient modification☆17Updated last year
- [ICML 2024] Official implementation for "Predictive Dynamic Fusion."☆70Updated last year
- The repo for "MMPareto: Boosting Multimodal Learning with Innocent Unimodal Assistance", ICML 2024☆54Updated last year
- ☆25Updated 3 months ago
- [ICML 2023] Provable Dynamic Fusion for Low-Quality Multimodal Data☆116Updated 7 months ago
- Multimodal Prompting with Missing Modalities for Visual Recognition, CVPR'23☆226Updated 2 years ago
- Code for the paper 'Dynamic Multimodal Fusion'☆122Updated 2 years ago
- A curated list of balanced multimodal learning methods.☆154Updated this week
- The repo for "Balanced Multimodal Learning via On-the-fly Gradient Modulation", CVPR 2022 (ORAL)☆308Updated 4 months ago
- ☆39Updated 5 months ago
- A novel cross-modal decoupling and alignment framework for multimodal representation learning.☆44Updated last month
- [CVPR 2024] Official PyTorch Code for "PromptKD: Unsupervised Prompt Distillation for Vision-Language Models"☆347Updated last month
- Code for UniS-MMC: Multimodal Classification via Unimodality-supervised Multimodal Contrastive Learning (ACL 2023)☆38Updated last year
- Code for the paper Visual Explanations of Image–Text Representations via Multi-Modal Information Bottleneck Attribution☆64Updated last year
- An official implementation of "Distribution-Consistent Modal Recovering for Incomplete Multimodal Learning" in PyTorch. (ICCV 2023)☆34Updated 2 years ago
- [CVPR2024] Continual-MAE: Adaptive Distribution Masked Autoencoders for Continual Test-Time Adaptation☆19Updated last year
- Code for "Leveraging Knowledge of Modality Experts for Incomplete Multimodal Learning" accepted by ACM Multimedia 2024☆43Updated last year
- The official code repository of ShaSpec model from CVPR 2023 [paper](https://arxiv.org/pdf/2307.14126) "Multi-modal Learning with Missing…☆88Updated 9 months ago
- Source code for the paper "Long-Tail Learning with Foundation Model: Heavy Fine-Tuning Hurts" (ICML 2024)☆100Updated last year
- MAE for CIFAR,由于可用资源有限,我们仅在 cifar10 上测试模型。我们主要想重现这样的结果:使用 MAE 预训练 ViT 可以比直接使用标签进行监督学习训练获得更好的结果。这应该是自我监督学习比监 督学习更有效的数据的证据。☆81Updated 2 years ago
- This is the official implementation of "Fuzzy Multimodal Learning for Trusted Cross-modal Retrieval" (CVPR 2025)☆39Updated 2 months ago