rinnakk / prefix-tuning-gptLinks
Example code for prefix-tuning GPT/GPT-NeoX models and for inference with trained prefixes
☆13Updated 2 years ago
Alternatives and similar repositories for prefix-tuning-gpt
Users that are interested in prefix-tuning-gpt are comparing it to the libraries listed below
Sorting:
- ☆46Updated 3 years ago
- Checkpointable dataset utilities for foundation model training☆32Updated last year
- ☆29Updated 3 years ago
- A Benchmark for Robust, Multi-evidence, Multi-answer Question Answering☆16Updated 2 years ago
- Observe the slow deterioration of my mental sanity in the github commit history☆12Updated 2 years ago
- ☆49Updated last year
- ☆11Updated 3 years ago
- Codes to pre-train Japanese T5 models☆41Updated 3 years ago
- Source code for the GPT-2 story generation models in the EMNLP 2020 paper "STORIUM: A Dataset and Evaluation Platform for Human-in-the-Lo…☆39Updated last year
- ☆43Updated 3 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Codebase for public release of the plug-and-blend framework.☆23Updated 3 years ago
- KETOD Knowledge-Enriched Task-Oriented Dialogue☆32Updated 2 years ago
- Do Multilingual Language Models Think Better in English?☆42Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 2 years ago
- ☆17Updated 7 months ago
- Lightblue LLM Eval Framework: tengu, elyza100, ja-mtbench, rakuda☆13Updated this week
- Project of llm evaluation to Japanese tasks☆85Updated this week
- M2D2: A Massively Multi-domain Language Modeling Dataset (EMNLP 2022) by Machel Reid, Victor Zhong, Suchin Gururangan, Luke Zettlemoyer☆54Updated 2 years ago
- ☆28Updated 3 months ago
- Convenient Text-to-Text Training for Transformers☆19Updated 3 years ago
- Utility scripts for preprocessing Wikipedia texts for NLP☆77Updated last year
- Support Continual pre-training & Instruction Tuning forked from llama-recipes☆32Updated last year
- IntructIR, a novel benchmark specifically designed to evaluate the instruction following ability in information retrieval models. Our foc…☆32Updated last year
- Code and dataset "ZEST" from "Learning from task descriptions", Weller et al, EMNLP 2020☆17Updated 4 years ago
- BLOOM+1: Adapting BLOOM model to support a new unseen language☆73Updated last year
- ☆15Updated 3 years ago
- COMET-ATOMIC ja☆30Updated last year
- The official implementation of "Distilling Relation Embeddings from Pre-trained Language Models, EMNLP 2021 main conference", a high-qual…☆47Updated 7 months ago
- A simple implementation of SimCSE☆77Updated 2 years ago