JoaoLages / RATransformersLinks
RATransformers π- Make your transformer (like BERT, RoBERTa, GPT-2 and T5) Relation Aware!
β42Updated 3 years ago
Alternatives and similar repositories for RATransformers
Users that are interested in RATransformers are comparing it to the libraries listed below
Sorting:
- Open source library for few shot NLPβ78Updated 2 years ago
- Embedding Recycling for Language modelsβ38Updated 2 years ago
- β21Updated 4 years ago
- β75Updated 4 years ago
- Ranking of fine-tuned HF models as base models.β36Updated 4 months ago
- This repository contains the code for the paper 'PARM: Paragraph Aggregation Retrieval Model for Dense Document-to-Document Retrieval' puβ¦β41Updated 4 years ago
- Code & data for EMNLP 2020 paper "MOCHA: A Dataset for Training and Evaluating Reading Comprehension Metrics".β16Updated 3 years ago
- Semantically Structured Sentence Embeddingsβ71Updated last year
- β22Updated 3 years ago
- Source code and data for Like a Good Nearest Neighborβ30Updated last year
- Starbucks: Improved Training for 2D Matryoshka Embeddingsβ22Updated 7 months ago
- β59Updated 4 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.β96Updated 2 years ago
- Official code and model checkpoints for our EMNLP 2022 paper "RankGen - Improving Text Generation with Large Ranking Models" (https://arxβ¦β138Updated 2 years ago
- β54Updated 3 years ago
- Code release for Type-Aware Bi-Encoders for Open-Domain Entity Retrievalβ19Updated 3 years ago
- Apps built using Inspired Cognition's Critique.β57Updated 2 years ago
- No Parameter Left Behind: How Distillation and Model Size Affect Zero-Shot Retrievalβ29Updated 3 years ago
- SMASHED is a toolkit designed to apply transformations to samples in datasets, such as fields extraction, tokenization, prompting, batchiβ¦β35Updated last year
- SPRINT Toolkit helps you evaluate diverse neural sparse models easily using a single click on any IR dataset.β47Updated 2 years ago
- Tutorial to pretrain & fine-tune a π€ Flax T5 model on a TPUv3-8 with GCPβ58Updated 3 years ago
- β44Updated last year
- π€ Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.β81Updated 3 years ago
- Data Programming by Demonstration (DPBD) for Document Classificationβ35Updated 4 years ago
- Ensembling Hugging Face transformers made easyβ61Updated 3 years ago
- Code associated with the paper "Entropy-based Attention Regularization Frees Unintended Bias Mitigation from Lists"β50Updated 3 years ago
- As good as new. How to successfully recycle English GPT-2 to make models for other languages (ACL Findings 2021)β48Updated 4 years ago
- Training T5 to perform numerical reasoning.β23Updated 4 years ago
- Shared code for training sentence embeddings with Flax / JAXβ28Updated 4 years ago
- Hugging Face RoBERTa with Flash Attention 2β24Updated 4 months ago