GiovanniTRA / MultimodalNeuralDatabasesLinks
This repository contains the code for the perspective paper "Multimodal Neural Databases" accepted at SIGIR 2023.
☆18Updated 7 months ago
Alternatives and similar repositories for MultimodalNeuralDatabases
Users that are interested in MultimodalNeuralDatabases are comparing it to the libraries listed below
Sorting:
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆48Updated last year
- Official repository with code and data accompanying the NAACL 2021 paper "Hurdles to Progress in Long-form Question Answering" (https://a…☆46Updated 2 years ago
- Code for "Finetuning Pretrained Transformers into Variational Autoencoders"☆39Updated 3 years ago
- Repo for ICML23 "Why do Nearest Neighbor Language Models Work?"☆58Updated 2 years ago
- A library for parameter-efficient and composable transfer learning for NLP with sparse fine-tunings.☆74Updated 11 months ago
- Official Repository of Pretraining Without Attention (BiGS), BiGS is the first model to achieve BERT-level transfer learning on the GLUE …☆113Updated last year
- Efficient Transformers with Dynamic Token Pooling☆63Updated 2 years ago
- This is the official PyTorch repo for "UNIREX: A Unified Learning Framework for Language Model Rationale Extraction" (ICML 2022).☆26Updated 2 years ago
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆62Updated 3 years ago
- The official repository for our paper "The Neural Data Router: Adaptive Control Flow in Transformers Improves Systematic Generalization".☆33Updated last month
- Standalone Product Key Memory module in Pytorch - for augmenting Transformer models☆82Updated 11 months ago
- [ACL 2023]: Training Trajectories of Language Models Across Scales https://arxiv.org/pdf/2212.09803.pdf☆24Updated last year
- [NeurIPS 2023 Main Track] This is the repository for the paper titled "Don’t Stop Pretraining? Make Prompt-based Fine-tuning Powerful Lea…☆74Updated last year
- Hierarchical Attention Transformers (HAT)☆56Updated last year
- No Parameters Left Behind: Sensitivity Guided Adaptive Learning Rate for Training Large Transformer Models (ICLR 2022)☆30Updated 3 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆50Updated 3 years ago
- The official repository for our paper "The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers". We s…☆67Updated 2 years ago
- ☆46Updated 3 years ago
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆58Updated last year
- ☆13Updated 2 years ago
- Embedding Recycling for Language models☆39Updated 2 years ago
- Implementation of Memformer, a Memory-augmented Transformer, in Pytorch☆119Updated 4 years ago
- Official code and model checkpoints for our EMNLP 2022 paper "RankGen - Improving Text Generation with Large Ranking Models" (https://arx…☆136Updated last year
- EMNLP 2022: Finding Dataset Shortcuts with Grammar Induction https://arxiv.org/abs/2210.11560☆58Updated 4 months ago
- ☆97Updated last year
- Evaluation pipeline for the BabyLM Challenge 2023.☆76Updated last year
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- ☆84Updated 11 months ago
- ☆51Updated 2 years ago
- Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data☆57Updated 3 years ago