kyegomez / NeVALinks
The open source implementation of "NeVA: NeMo Vision and Language Assistant"
β17Updated last year
Alternatives and similar repositories for NeVA
Users that are interested in NeVA are comparing it to the libraries listed below
Sorting:
- Unofficial implementation and experiments related to Set-of-Mark (SoM) ποΈβ87Updated last year
- Documentation, notes, links, etc for streams.β83Updated last year
- Evaluate the performance of computer vision models and prompts for zero-shot models (Grounding DINO, CLIP, BLIP, DINOv2, ImageBind, modelβ¦β36Updated last year
- EdgeSAM model for use with Autodistill.β27Updated last year
- This is the repository for the Photorealistic Unreal Graphics (PUG) datasets for representation learning.β237Updated last year
- Finetune any model on HF in less than 30 secondsβ57Updated 2 weeks ago
- ClickDiffusion: Harnessing LLMs for Interactive Precise Image Editingβ69Updated last year
- The Next Generation Multi-Modality Superintelligenceβ70Updated 11 months ago
- β14Updated last year
- β69Updated last year
- β54Updated last year
- (WACV 2025 - Oral) Vision-language conversation in 10 languages including English, Chinese, French, Spanish, Russian, Japanese, Arabic, Hβ¦β83Updated this week
- Multi-model video-to-text by combining embeddings from Flan-T5 + CLIP + Whisper + SceneGraph. The 'backbone LLM' is pre-trained from scraβ¦β52Updated 2 years ago
- Implementation of the text to video model LUMIERE from the paper: "A Space-Time Diffusion Model for Video Generation" by Google Researchβ51Updated 6 months ago
- Use Grounding DINO, Segment Anything, and GPT-4V to label images with segmentation masks for use in training smaller, fine-tuned models.β66Updated last year
- Implementation of VisionLLaMA from the paper: "VisionLLaMA: A Unified LLaMA Interface for Vision Tasks" in PyTorch and Zetaβ16Updated 8 months ago
- Internet Explorer explores the web in a self-supervised manner to progressively find relevant examples that improve performance on a desiβ¦β163Updated 2 years ago
- Implementation of the paper: "BRAVE : Broadening the visual encoding of vision-language models"β27Updated this week
- A multi-modal AI Model that can generate high quality novel videos with text, images, or video clips.β64Updated last year
- Visual RAG using less than 300 lines of code.β28Updated last year
- A simple package for leveraging Falcon 180B and the HF ecosystem's tools, including training/inference scripts, safetensors, integrationsβ¦β12Updated last year
- [IJCAI'23] Complete Instances Mining for Weakly Supervised Instance Segmentationβ37Updated last year
- Use Grounding DINO, Segment Anything, and CLIP to label objects in images.β31Updated last year
- Cerule - A Tiny Mighty Vision Modelβ66Updated 11 months ago
- Implementation of the premier Text to Video model from OpenAIβ56Updated 8 months ago
- An EXA-Scale repository of Multi-Modality AI resources from papers and models, to foundational libraries!β40Updated last year
- β59Updated last year
- MetaCLIP module for use with Autodistill.β21Updated last year
- Repository for the paper: "TiC-CLIP: Continual Training of CLIP Models".β102Updated last year
- Use Segment Anything 2, grounded with Florence-2, to auto-label data for use in training vision models.β126Updated last year