Geotrend-research / smaller-transformers
Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.
β101Updated 2 years ago
Related projects β
Alternatives and complementary repositories for smaller-transformers
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.β92Updated last year
- Tutorial to pretrain & fine-tune a π€ Flax T5 model on a TPUv3-8 with GCPβ58Updated 2 years ago
- Python-based implementation of the Translate-Align-Retrieve method to automatically translate the SQuAD Dataset to Spanish.β59Updated last year
- β73Updated 3 years ago
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scaleβ153Updated 11 months ago
- Zero-shot Transfer Learning from English to Arabicβ29Updated 2 years ago
- Mr. TyDi is a multi-lingual benchmark dataset built on TyDi, covering eleven typologically diverse languages.β72Updated 2 years ago
- Shared code for training sentence embeddings with Flax / JAXβ27Updated 3 years ago
- Fine-tune transformers with pytorch-lightningβ44Updated 2 years ago
- β95Updated last year
- Code and datasets of "Multilingual Extractive Reading Comprehension by Runtime Machine Translation"β39Updated 5 years ago
- This repository contains the code for the paper 'PARM: Paragraph Aggregation Retrieval Model for Dense Document-to-Document Retrieval' puβ¦β40Updated 2 years ago
- State of the art Semantic Sentence Embeddingsβ98Updated 2 years ago
- β86Updated 2 years ago
- β83Updated 2 months ago
- This repository contains the code for "Generating Datasets with Pretrained Language Models".β187Updated 3 years ago
- Tools for managing datasets for governance and training.β78Updated 3 weeks ago
- A package for fine-tuning Transformers with TPUs, written in Tensorflow2.0+β37Updated 3 years ago
- β16Updated last year
- Pytorch Implementation of EncT5: Fine-tuning T5 Encoder for Non-autoregressive Tasksβ63Updated 2 years ago
- A tiny BERT for low-resource monolingual modelsβ29Updated last month
- Distillation of BERT model with catalyst frameworkβ75Updated last year
- π οΈ Tools for Transformers compression using PyTorch Lightning β‘β79Updated last week
- β67Updated 3 years ago
- Accelerated NLP pipelines for fast inference on CPU. Built with Transformers and ONNX runtime.β126Updated 3 years ago
- A repository for our AAAI-2020 Cross-lingual-NER paper. Code will be updated shortly.β46Updated last year
- Implementation of Marge, Pre-training via Paraphrasing, in Pytorchβ75Updated 3 years ago
- Dual Encoders for State-of-the-art Natural Language Processing.β61Updated 2 years ago
- β21Updated 3 years ago
- BERT, RoBERTa fine-tuning over SQuAD Dataset using pytorch-lightningβ‘οΈ, π€-transformers & π€-nlp.β36Updated last year