ShilinSun / mxai_reviewLinks
☆10Updated 6 months ago
Alternatives and similar repositories for mxai_review
Users that are interested in mxai_review are comparing it to the libraries listed below
Sorting:
- MuCR is a benchmark designed to evaluate Multimodal Large Language Models' (MLLMs) ability to discern causal links across modalities☆15Updated last month
- The official GitHub page for paper "NegativePrompt: Leveraging Psychology for Large Language Models Enhancement via Negative Emotional St…☆23Updated last year
- ☆17Updated 11 months ago
- Parameter-Efficient Fine-Tuning for Foundation Models☆75Updated 3 months ago
- ROUTE: Robust Multitask Tuning and Collaboration for Text-to-SQL (ICLR 2025 Pytorch Code)☆15Updated 2 months ago
- The benchmark and datasets of the ICML 2024 paper "VisionGraph: Leveraging Large Multimodal Models for Graph Theory Problems in Visual C…☆14Updated last year
- ☆21Updated 8 months ago
- Repo of FocusedAD☆13Updated 3 months ago
- [ICML 2025] Code for "R2-T2: Re-Routing in Test-Time for Multimodal Mixture-of-Experts"☆15Updated 4 months ago
- [CVPR 2025] Offical implementation of the paper "Skip Tuning: Pre-trained Vision-Language Models are Effective and Efficient Adapters The…☆18Updated 4 months ago
- RewardAnything: Generalizable Principle-Following Reward Models☆25Updated last month
- Source code of paper: A Stronger Mixture of Low-Rank Experts for Fine-Tuning Foundation Models. (ICML 2025)☆27Updated 3 months ago
- The official repository of "SmartAgent: Chain-of-User-Thought for Embodied Personalized Agent in Cyber World".☆28Updated 3 months ago
- Sparse Autoencoders Learn Monosemantic Features in Vision-Language Models☆24Updated 3 months ago
- WorldSense: Evaluating Real-world Omnimodal Understanding for Multimodal LLMs☆26Updated 2 months ago
- Code for paper "Unraveling Cross-Modality Knowledge Conflicts in Large Vision-Language Models."☆42Updated 9 months ago
- ☆13Updated 7 months ago
- ☆11Updated 3 months ago
- codes for Efficient Test-Time Scaling via Self-Calibration☆14Updated 4 months ago
- [CVPR 2025] Official PyTorch Code for "MMRL: Multi-Modal Representation Learning for Vision-Language Models" and its extension "MMRL++: P…☆57Updated 3 weeks ago
- ☆21Updated 3 months ago
- [ACL 2023] Code for paper “Tailoring Instructions to Student’s Learning Levels Boosts Knowledge Distillation”(https://arxiv.org/abs/2305.…☆38Updated 2 years ago
- ☆18Updated 4 months ago
- Code for "CREAM: Consistency Regularized Self-Rewarding Language Models", ICLR 2025.☆22Updated 5 months ago
- ☆43Updated 2 months ago
- Do Vision and Language Models Share Concepts? A Vector Space Alignment Study☆15Updated 7 months ago
- ☆19Updated 2 months ago
- Codes for our paper "AgentMonitor: A Plug-and-Play Framework for Predictive and Secure Multi-Agent Systems"☆10Updated 7 months ago
- The official repo for "VisualWebInstruct: Scaling up Multimodal Instruction Data through Web Search"☆25Updated 2 months ago
- MultiModN – Multimodal, Multi-Task, Interpretable Modular Networks (NeurIPS 2023)☆33Updated last year