nengelmann / Fuyu-8B---ExplorationLinks
Exploration of the multi modal fuyu-8b model of Adept. π€ π
β27Updated 2 years ago
Alternatives and similar repositories for Fuyu-8B---Exploration
Users that are interested in Fuyu-8B---Exploration are comparing it to the libraries listed below
Sorting:
- imagetokenizer is a python package, helps you encoder visuals and generate visuals token ids from codebook, supports both image and videoβ¦β37Updated last year
- Our 2nd-gen LMMβ34Updated last year
- A Framework for Decoupling and Assessing the Capabilities of VLMsβ43Updated last year
- In-Context Alignment: Chat with Vanilla Language Models Before Fine-Tuningβ35Updated 2 years ago
- Web2Code: A Large-scale Webpage-to-Code Dataset and Evaluation Framework for Multimodal LLMsβ92Updated last year
- β75Updated last year
- A Simple MLLM Surpassed QwenVL-Max with OpenSource Data Only in 14B LLM.β38Updated last year
- An End-to-End Model with Adaptive Filtering for Retrieval-Augmented Generationβ15Updated last year
- Simple Implementation of TinyGPTV in super simple Zeta lego blocksβ15Updated last year
- β29Updated last year
- XVERSE-MoE-A36B: A multilingual large language model developed by XVERSE Technology Inc.β38Updated last year
- β15Updated last year
- [NAACL 2025] Representing Rule-based Chatbots with Transformersβ22Updated 9 months ago
- Code for Paper: Harnessing Webpage Uis For Text Rich Visual Understandingβ53Updated 11 months ago
- A multimodal large-scale model, which performs close to the closed-source Qwen-VL-PLUS on many datasets and significantly surpasses the pβ¦β14Updated last year
- [ACL2025 Findings] Benchmarking Multihop Multimodal Internet Agentsβ47Updated 8 months ago
- A tiny, didactical implementation of LLAMA 3β42Updated 11 months ago
- An efficient multi-modal instruction-following data synthesis tool and the official implementation of Oasis https://arxiv.org/abs/2503.08β¦β32Updated 5 months ago
- A minimal implementation of LLaVA-style VLM with interleaved image & text & video processing ability.β96Updated 11 months ago
- Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, modelβ¦β37Updated 2 years ago
- β57Updated last year
- Pytorch implementation of HyperLLaVA: Dynamic Visual and Language Expert Tuning for Multimodal Large Language Modelsβ28Updated last year
- Tools for content datamining and NLP at scaleβ44Updated last year
- β28Updated 3 months ago
- β28Updated last year
- EfficientSAM + YOLO World base model for use with Autodistill.β10Updated last year
- Fast LLM Training CodeBase With dynamic strategy choosing [Deepspeed+Megatron+FlashAttention+CudaFusionKernel+Compiler];β41Updated last year
- Official Pytorch Implementation of Self-emerging Token Labelingβ35Updated last year
- Empirical Study Towards Building An Effective Multi-Modal Large Language Modelβ22Updated 2 years ago
- helper functions for processing and integrating visual language information with Qwen-VL Series Modelβ16Updated last year