chrishokamp / zero-shot-ner-fine-tuningLinks
zero shot NER fine tuning
☆13Updated 6 months ago
Alternatives and similar repositories for zero-shot-ner-fine-tuning
Users that are interested in zero-shot-ner-fine-tuning are comparing it to the libraries listed below
Sorting:
- A tiny BERT for low-resource monolingual models☆31Updated last year
- Data and evaluation code for the paper WikiNEuRal: Combined Neural and Knowledge-based Silver Data Creation for Multilingual NER (EMNLP 2…☆68Updated 2 years ago
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 3 years ago
- This repository contains the code for the paper 'PARM: Paragraph Aggregation Retrieval Model for Dense Document-to-Document Retrieval' pu…☆41Updated 3 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆95Updated 2 years ago
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆63Updated 3 weeks ago
- This repo contains a set of neural transducer, e.g. sequence-to-sequence model, focusing on character-level tasks.☆76Updated 2 years ago
- PyTorch-IE: State-of-the-art Information Extraction in PyTorch☆78Updated this week
- Implementation of Z-BERT-A: a zero-shot pipeline for unknown intent detection.☆43Updated 2 years ago
- Repository for the paper "MultiNERD: A Multilingual, Multi-Genre and Fine-Grained Dataset for Named Entity Recognition (and Disambiguatio…☆45Updated last year
- Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.☆105Updated 3 years ago
- ☆34Updated 4 years ago
- Text Extraction Formulation + Feedback Loop for state-of-the-art WSD (EMNLP 2021)☆53Updated 3 years ago
- No Parameter Left Behind: How Distillation and Model Size Affect Zero-Shot Retrieval☆29Updated 3 years ago
- Dataset from the paper "Mintaka: A Complex, Natural, and Multilingual Dataset for End-to-End Question Answering" (COLING 2022)☆114Updated 2 years ago
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆27Updated last year
- Source code for ASRU 2019 paper "Adapting Pretrained Transformer to Lattices for Spoken Language Understanding"☆11Updated 5 years ago
- zero-vocab or low-vocab embeddings☆18Updated 3 years ago
- Vocabulary Trimming (VT) is a model compression technique, which reduces a multilingual LM vocabulary to a target language by deleting ir…☆46Updated 11 months ago
- Fast whitespace correction with Transformers☆17Updated last month
- Code for WECHSEL: Effective initialization of subword embeddings for cross-lingual transfer of monolingual language models.☆84Updated last year
- Experiments for XLM-V Transformers Integeration☆13Updated 2 years ago
- Personal information identification standard☆20Updated last year
- Pre-train Static Word Embeddings☆85Updated 2 weeks ago
- Semantically Structured Sentence Embeddings☆68Updated 11 months ago
- ☆51Updated 2 years ago
- LTG-Bert☆33Updated last year
- Tools for managing datasets for governance and training.☆83Updated this week
- An asynchronous concurrent pipeline for classifying Common Crawl based on fastText's pipeline.☆86Updated 4 years ago
- BLOOM+1: Adapting BLOOM model to support a new unseen language☆73Updated last year