rinnakk / prefix-tuning-gptLinks
Example code for prefix-tuning GPT/GPT-NeoX models and for inference with trained prefixes
☆13Updated 2 years ago
Alternatives and similar repositories for prefix-tuning-gpt
Users that are interested in prefix-tuning-gpt are comparing it to the libraries listed below
Sorting:
- A Benchmark for Robust, Multi-evidence, Multi-answer Question Answering☆17Updated 2 years ago
- ☆46Updated 3 years ago
- ☆30Updated 3 years ago
- Lightblue LLM Eval Framework: tengu, elyza100, ja-mtbench, rakuda☆18Updated 3 weeks ago
- ☆11Updated 4 years ago
- ☆43Updated 4 years ago
- Checkpointable dataset utilities for foundation model training☆32Updated last year
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆57Updated 3 years ago
- Support Continual pre-training & Instruction Tuning forked from llama-recipes☆33Updated last year
- Observe the slow deterioration of my mental sanity in the github commit history☆12Updated 2 years ago
- Implementation of autoregressive language model using improved Transformer and DeepSpeed pipeline parallelism.☆32Updated 3 years ago
- ☆14Updated 3 years ago
- ☆43Updated last year
- Project of llm evaluation to Japanese tasks☆90Updated last month
- ☆26Updated 8 months ago
- ☆44Updated last year
- KETOD Knowledge-Enriched Task-Oriented Dialogue☆32Updated 2 years ago
- BLOOM+1: Adapting BLOOM model to support a new unseen language☆74Updated last year
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆35Updated 2 years ago
- Convenient Text-to-Text Training for Transformers☆19Updated 3 years ago
- [COLM 2024] Early Weight Averaging meets High Learning Rates for LLM Pre-training☆18Updated last year
- Helper scripts and notes that were used while porting various nlp models☆48Updated 3 years ago
- ☆15Updated 3 years ago
- COMET-ATOMIC ja☆31Updated last year
- DIRECT: Direct and Indirect REsponses in Conversational Text Corpus☆17Updated 4 years ago
- GPT-jax based on the official huggingface library☆13Updated 4 years ago
- Codes to pre-train Japanese T5 models☆40Updated 4 years ago
- ☆28Updated 7 months ago
- Emotion-Aware Dialogue Response Generation by Multi-Task Learning☆13Updated 3 years ago
- Code associated with the "Data Augmentation using Pre-trained Transformer Models" paper☆51Updated 2 years ago