philschmid / optimum-transformers-optimizations
☆30Updated 2 years ago
Related projects ⓘ
Alternatives and complementary repositories for optimum-transformers-optimizations
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆92Updated last year
- Using short models to classify long texts☆20Updated last year
- Anh - LAION's multilingual assistant datasets and models☆27Updated last year
- ☆21Updated 3 years ago
- exBERT on Transformers🤗☆10Updated 3 years ago
- Code for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models.☆75Updated 2 months ago
- **ARCHIVED** Filesystem interface to 🤗 Hub☆56Updated last year
- ☆19Updated 3 years ago
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆44Updated last year
- A Framework aims to wisely initialize unseen subword embeddings in PLMs for efficient large-scale continued pretraining☆12Updated 11 months ago
- Ensembling Hugging Face transformers made easy☆62Updated last year
- Scripts to convert datasets from various sources to Hugging Face Datasets.☆57Updated 2 years ago
- Observe the slow deterioration of my mental sanity in the github commit history☆13Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- Applying "Load What You Need: Smaller Versions of Multilingual BERT" to LaBSE☆18Updated 3 years ago
- Mr. TyDi is a multi-lingual benchmark dataset built on TyDi, covering eleven typologically diverse languages.☆72Updated 2 years ago
- Personal information identification standard☆19Updated 9 months ago
- Ranking of fine-tuned HF models as base models.☆35Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Updated 2 months ago
- Embedding Recycling for Language models☆38Updated last year
- Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.☆101Updated 2 years ago
- ☆28Updated last year
- Consists of the largest (10K) human annotated code-switched semantic parsing dataset & 170K generated utterance using the CST5 augmentati…☆33Updated last year
- ☆20Updated last year
- ☆15Updated 3 months ago
- Train 🤗transformers with DeepSpeed: ZeRO-2, ZeRO-3☆21Updated 3 years ago
- ☆29Updated last year
- This repository contains the code for the paper 'PARM: Paragraph Aggregation Retrieval Model for Dense Document-to-Document Retrieval' pu…☆40Updated 2 years ago