google-research / scenic
Scenic: A Jax Library for Computer Vision Research and Beyond
☆3,385Updated this week
Alternatives and similar repositories for scenic:
Users that are interested in scenic are comparing it to the libraries listed below
- Official DeiT repository☆4,114Updated 10 months ago
- Official codebase used to develop Vision Transformer, SigLIP, MLP-Mixer, LiT and more.☆2,516Updated 3 weeks ago
- VISSL is FAIR's library of extensible, modular and scalable components for SOTA Self-Supervised Learning with images.☆3,264Updated 10 months ago
- PyTorch code for Vision Transformers training with the Self-Supervised learning method DINO☆6,525Updated 6 months ago
- Code release for ConvNeXt model☆5,845Updated 2 years ago
- Grounded Language-Image Pre-training☆2,297Updated 11 months ago
- ☆10,780Updated last month
- The official pytorch implementation of our paper "Is Space-Time Attention All You Need for Video Understanding?"☆1,601Updated 9 months ago
- EVA Series: Visual Representation Fantasies from BAAI☆2,381Updated 5 months ago
- A deep learning library for video understanding research.☆3,365Updated last month
- An ultimately comprehensive paper list of Vision Transformer/Attention, including papers, codes, and related websites☆4,716Updated 5 months ago
- TorchMultimodal is a PyTorch library for training state-of-the-art multimodal multi-task models at scale.☆1,519Updated this week
- Code release for "Masked-attention Mask Transformer for Universal Image Segmentation"☆2,655Updated 5 months ago
- Collection of common code that's shared among different research projects in FAIR computer vision team.☆2,057Updated last month
- CVNets: A library for training computer vision networks☆1,818Updated last year
- This is an official implementation for "Video Swin Transformers".☆1,478Updated last year
- OpenMMLab Self-Supervised Learning Toolbox and Benchmark☆3,224Updated last year
- SimCLRv2 - Big Self-Supervised Models are Strong Semi-Supervised Learners☆4,173Updated last year
- Implementation of CoCa, Contrastive Captioners are Image-Text Foundation Models, in Pytorch☆1,092Updated last year
- PyTorch implementation of MoCo v3 https//arxiv.org/abs/2104.02057☆1,235Updated 3 years ago
- This is a collection of our NAS and Vision Transformer work.☆1,707Updated 5 months ago
- PyTorch implementation of MoCo: https://arxiv.org/abs/1911.05722☆4,855Updated 3 weeks ago
- Hiera: A fast, powerful, and simple hierarchical vision transformer.☆940Updated 10 months ago
- Code release for ConvNeXt V2 model☆1,584Updated 5 months ago
- [CVPR 2021] Official PyTorch implementation for Transformer Interpretability Beyond Attention Visualization, a novel method to visualize …☆1,827Updated 11 months ago
- [NeurIPS 2022 Spotlight] VideoMAE: Masked Autoencoders are Data-Efficient Learners for Self-Supervised Video Pre-Training☆1,421Updated last year
- Pytorch reimplementation of the Vision Transformer (An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale)☆1,978Updated 2 years ago
- [CVPR 2023 Highlight] InternImage: Exploring Large-Scale Vision Foundation Models with Deformable Convolutions☆2,575Updated this week
- Recent Transformer-based CV and related works.☆1,326Updated last year
- FFCV: Fast Forward Computer Vision (and other ML workloads!)☆2,880Updated 7 months ago