shan18 / Perceiver-Resampler-XAttn-Captioning
Generating Captions via Perceiver-Resampler Cross-Attention Networks
☆14Updated last year
Related projects ⓘ
Alternatives and complementary repositories for Perceiver-Resampler-XAttn-Captioning
- ☆38Updated 3 months ago
- ☆50Updated last year
- Code used for the creation of OBELICS, an open, massive and curated collection of interleaved image-text web documents, containing 141M d…☆188Updated 2 months ago
- ☆27Updated this week
- Utilities for Training Very Large Models☆56Updated last month
- [Under Review] Official PyTorch implementation code for realizing the technical part of Phantom of Latent representing equipped with enla…☆45Updated last month
- ☆57Updated 8 months ago
- ☆50Updated 2 weeks ago
- Code repository for the public reproduction of the language modelling experiments on "MatFormer: Nested Transformer for Elastic Inference…☆18Updated last year
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆43Updated last month
- ☆63Updated 4 months ago
- Code for NOLA, an implementation of "nola: Compressing LoRA using Linear Combination of Random Basis"☆49Updated 2 months ago
- Code and pretrained models for the paper: "MatMamba: A Matryoshka State Space Model"☆50Updated last month
- ☆21Updated 7 months ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆79Updated last year
- Code for T-MARS data filtering☆35Updated last year
- Language models scale reliably with over-training and on downstream tasks☆94Updated 7 months ago
- Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount…☆49Updated last year
- M4 experiment logbook☆56Updated last year
- Implementation of the paper: "BRAVE : Broadening the visual encoding of vision-language models"☆21Updated this week
- The open source implementation of the model from "Scaling Vision Transformers to 22 Billion Parameters"☆25Updated this week
- [NeurIPS 2024] Goldfish Loss: Mitigating Memorization in Generative LLMs☆74Updated 3 weeks ago
- code for paper "Accessing higher dimensions for unsupervised word translation"☆21Updated last year
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆64Updated last year
- Repository for the paper: "TiC-CLIP: Continual Training of CLIP Models".☆93Updated 5 months ago
- ☆45Updated 9 months ago
- Implementation of some personal helper functions for Einops, my most favorite tensor manipulation library ❤️☆52Updated last year
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- ☆29Updated this week
- Experimental scripts for researching data adaptive learning rate scheduling.☆23Updated last year