Rishit-dagli / Compositional-AttentionLinks
An implementation of Compositional Attention: Disentangling Search and Retrieval by MILA
☆14Updated 3 years ago
Alternatives and similar repositories for Compositional-Attention
Users that are interested in Compositional-Attention are comparing it to the libraries listed below
Sorting:
- PyTorch Implementation of the paper "MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training"☆24Updated last week
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆50Updated 3 years ago
- Implementation of Metaformer, but in an autoregressive manner☆26Updated 3 years ago
- Directed masked autoencoders☆14Updated 2 years ago
- A python library for highly configurable transformers - easing model architecture search and experimentation.☆49Updated 4 years ago
- code for paper "Accessing higher dimensions for unsupervised word translation"☆22Updated 2 years ago
- Applying Reinforcement Learning from Human Feedback to language models to teach them to write short story responses to writing prompts.☆14Updated 3 years ago
- An attempt to merge ESBN with Transformers, to endow Transformers with the ability to emergently bind symbols☆16Updated 4 years ago
- Load any clip model with a standardized interface☆22Updated last month
- Google Research☆46Updated 3 years ago
- LoRA fine-tuned Stable Diffusion Deployment☆31Updated 2 years ago
- Large dataset storage format for Pytorch☆45Updated 4 years ago
- SMASHED is a toolkit designed to apply transformations to samples in datasets, such as fields extraction, tokenization, prompting, batchi…☆35Updated last year
- Ranking of fine-tuned HF models as base models.☆36Updated 2 months ago
- ☆21Updated 2 years ago
- ☆31Updated last week
- Benchmarking algorithms for assessing quality of data labeled by multiple annotators☆34Updated 2 years ago
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆39Updated 3 years ago
- ☆44Updated last year
- Companion Repo for the Vision Language Modelling YouTube series - https://bit.ly/3PsbsC2 - by Prithivi Da. Open to PRs and collaborations☆14Updated 3 years ago
- Utilities for Training Very Large Models☆58Updated last year
- Another attempt at a long-context / efficient transformer by me☆38Updated 3 years ago
- Implementation of Memory-Compressed Attention, from the paper "Generating Wikipedia By Summarizing Long Sequences"☆70Updated 2 years ago
- Embroid: Unsupervised Prediction Smoothing Can Improve Few-Shot Classification☆11Updated 2 years ago
- Visual Clustering: Clustering Plotted Data by Image Segmentation☆25Updated 9 months ago
- My explorations into editing the knowledge and memories of an attention network☆35Updated 2 years ago
- Experimental scripts for researching data adaptive learning rate scheduling.☆22Updated 2 years ago
- Code for our paper Resources and Evaluations for Multi-Distribution Dense Information Retrieval☆15Updated last year
- machine learning model performance metrics & charts with confidence intervals, optimized with numba to be fast☆16Updated 3 years ago