bhattbhavesh91 / few-shot-learning-using-gpt-neoLinks
Few Shot Learning using EleutherAI's GPT-Neo an Open-source version of GPT-3
☆18Updated 4 years ago
Alternatives and similar repositories for few-shot-learning-using-gpt-neo
Users that are interested in few-shot-learning-using-gpt-neo are comparing it to the libraries listed below
Sorting:
- Using short models to classify long texts☆21Updated 2 years ago
- URL downloader supporting checkpointing and continuous checksumming.☆19Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- Embedding Recycling for Language models☆38Updated 2 years ago
- A package for fine tuning of pretrained NLP transformers using Semi Supervised Learning☆14Updated 3 years ago
- GPT-jax based on the official huggingface library☆13Updated 4 years ago
- ☆12Updated 7 months ago
- Applying "Load What You Need: Smaller Versions of Multilingual BERT" to LaBSE☆18Updated 3 years ago
- ☆30Updated 2 years ago
- Implementation of stop sequencer for Huggingface Transformers☆16Updated 2 years ago
- ☆37Updated 2 years ago
- ☆32Updated 2 years ago
- Plug-and-play Search Interfaces with Pyserini and Hugging Face☆32Updated last year
- Developing tools to automatically analyze datasets☆74Updated 8 months ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated 2 years ago
- Consists of the largest (10K) human annotated code-switched semantic parsing dataset & 170K generated utterance using the CST5 augmentati…☆39Updated 2 years ago
- Entailment self-training☆25Updated 2 years ago
- 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.☆82Updated 3 years ago
- Convenient Text-to-Text Training for Transformers☆19Updated 3 years ago
- **ARCHIVED** Filesystem interface to 🤗 Hub☆58Updated 2 years ago
- All my experiments with the various transformers and various transformer frameworks available☆14Updated 4 years ago
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆27Updated last year
- A library for squeakily cleaning and filtering language datasets.☆47Updated 2 years ago
- ☆16Updated last year
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆61Updated last month
- My explorations into editing the knowledge and memories of an attention network☆35Updated 2 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Our open source implementation of MiniLMv2 (https://aclanthology.org/2021.findings-acl.188)☆61Updated 2 years ago
- Ensembling Hugging Face transformers made easy☆63Updated 2 years ago
- The collection of bulding blocks building fine-tunable metric learning models☆32Updated 3 months ago