hadarishav / Ruddit
This repo contains the dataset and description for Ruddit and its variants.
☆35Updated 2 years ago
Related projects: ⓘ
- Topic clustering library built on Transformer embeddings and cosine similarity metrics.Compatible with all BERT base transformers from hu…☆41Updated 3 years ago
- 1st solution☆37Updated last year
- Kaggle Tweet Sentiment Extraction Competition: 1st place solution (Dark of the Moon team)☆70Updated 2 years ago
- Early solution for Google AI4Code competition☆75Updated 2 years ago
- PyTorch – SMART: Robust and Efficient Fine-Tuning for Pre-trained Natural Language Models.☆59Updated 2 years ago
- ☆66Updated 2 years ago
- pyTorch implementation of Recurrence over BERT (RoBERT) based on this paper https://arxiv.org/abs/1910.10781 and comparison with pyTorch …☆77Updated last year
- State of the art Semantic Sentence Embeddings☆97Updated 2 years ago
- Implementation of Mixout with PyTorch☆73Updated last year
- [NAACL 2021] This is the code for our paper `Fine-Tuning Pre-trained Language Model with Weak Supervision: A Contrastive-Regularized Self…☆200Updated 2 years ago
- ☆16Updated 3 years ago
- A Light and Modular PyTorch NLP Project Template☆58Updated 4 years ago
- [NeurIPS 2021] COCO-LM: Correcting and Contrasting Text Sequences for Language Model Pretraining☆120Updated last year
- Self-supervised NER prototype - updated version (69 entity types - 17 broad entity groups). Uses pretrained BERT models with no fine tuni…☆80Updated 2 years ago
- ☆12Updated 2 years ago
- Interpreting Language Models with Contrastive Explanations (EMNLP 2022 Best Paper Honorable Mention)☆56Updated 2 years ago
- Efficient Attention for Long Sequence Processing☆84Updated 9 months ago
- Master thesis with code investigating methods for incorporating long-context reasoning in low-resource languages, without the need to pre…☆32Updated 3 years ago
- Code associated with the "Data Augmentation using Pre-trained Transformer Models" paper☆50Updated last year
- [EMNLP'21] Mirror-BERT: Converting Pretrained Language Models to universal text encoders without labels.☆75Updated 2 years ago
- Named Entity Recognition with Small Strongly Labeled and Large Weakly Labeled Data☆100Updated last year
- ☆40Updated 2 years ago
- ☆27Updated 3 years ago
- This is the solution for 2nd rank in Kaggle competition: Feedback Prize - Evaluating Student Writing.☆48Updated 2 years ago
- Long-context pretrained encoder-decoder models☆95Updated last year
- Pre-training BART in Flax on The Pile dataset☆20Updated 3 years ago
- On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines☆131Updated last year
- SNCSE: Contrastive Learning for Unsupervised Sentence Embedding with Soft Negative Samples☆73Updated 2 years ago
- ☆82Updated 3 years ago
- ☆42Updated 4 years ago