codefuse-ai / CodeFuse-MFT-VLM
☆36Updated 3 months ago
Alternatives and similar repositories for CodeFuse-MFT-VLM:
Users that are interested in CodeFuse-MFT-VLM are comparing it to the libraries listed below
- ☆78Updated 8 months ago
- A Simple MLLM Surpassed QwenVL-Max with OpenSource Data Only in 14B LLM.☆36Updated 4 months ago
- Valley is a cutting-edge multimodal large model designed to handle a variety of tasks involving text, images, and video data.☆197Updated last week
- Skywork-MoE: A Deep Dive into Training Techniques for Mixture-of-Experts Language Models☆128Updated 7 months ago
- LongQLoRA: Extent Context Length of LLMs Efficiently☆163Updated last year
- Delta-CoMe can achieve near loss-less 1-bit compressin which has been accepted by NeurIPS 2024☆52Updated 2 months ago
- Its an open source LLM based on MOE Structure.☆57Updated 6 months ago
- ☆167Updated last month
- SUS-Chat: Instruction tuning done right☆48Updated last year
- ☆56Updated last year
- Vision Search Assistant: Empower Vision-Language Models as Multimodal Search Engines☆109Updated 2 months ago
- Exploring Efficient Fine-Grained Perception of Multimodal Large Language Models☆56Updated 2 months ago
- ☆27Updated 5 months ago
- zero零训练llm调参☆31Updated last year
- [IJCAI 2024] CMMU: A Benchmark for Chinese Multi-modal Multi-type Question Understanding and Reasoning☆23Updated 11 months ago
- A Framework for Decoupling and Assessing the Capabilities of VLMs☆40Updated 7 months ago
- Multimodal chatbot with computer vision capabilities integrated☆100Updated 8 months ago
- XVERSE-MoE-A4.2B: A multilingual large language model developed by XVERSE Technology Inc.☆36Updated 8 months ago
- 基于baichuan-7b的开源多模态大语言模型☆73Updated last year
- 我们是第一个完全可商用的角色大模型。☆38Updated 5 months ago
- MuLan: Adapting Multilingual Diffusion Models for 110+ Languages (无需额外训练为任意扩散模型支持多语言能力)☆130Updated last week
- ☆67Updated last year
- Mixture-of-Experts (MoE) Language Model☆183Updated 4 months ago
- Empirical Study Towards Building An Effective Multi-Modal Large Language Model☆23Updated last year
- Multimodal Open-O1 (MO1) is designed to enhance the accuracy of inference models by utilizing a novel prompt-based approach. This tool wo…☆29Updated 4 months ago
- official code for "Fox: Focus Anywhere for Fine-grained Multi-page Document Understanding"☆136Updated 7 months ago
- DeepSpeed教程 & 示例注释 & 学习笔记 (大模型高效训练)☆146Updated last year
- A Toolkit for Running On-device Large Language Models (LLMs) in APP☆59Updated 6 months ago
- ☆32Updated 8 months ago
- A visuailzation tool to make deep understaning and easier debugging for RLHF training.☆129Updated 3 weeks ago