google-research-datasets / QAmeleon
QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning PaLM with only five examples per language. We use the synthetic data to finetune downstream QA models leading to improved accuracy in comparison to English-only and translation-based baselines.
☆34Updated last year
Related projects ⓘ
Alternatives and complementary repositories for QAmeleon
- Codes and files for the paper Are Emergent Abilities in Large Language Models just In-Context Learning☆34Updated 8 months ago
- ☆46Updated last week
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆23Updated 7 months ago
- QLoRA for Masked Language Modeling☆20Updated last year
- Embedding Recycling for Language models☆38Updated last year
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆44Updated last year
- Index of URLs to pdf files all over the internet and scripts☆21Updated last year
- My explorations into editing the knowledge and memories of an attention network☆34Updated last year
- This repository contains code for cleaning your training data of benchmark data to help combat data snooping.☆25Updated last year
- Code for the examples presented in the talk "Training a Llama in your backyard: fine-tuning very large models on consumer hardware" given…☆14Updated last year
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- ☆27Updated 5 months ago
- Code for NeurIPS LLM Efficiency Challenge☆54Updated 7 months ago
- Using short models to classify long texts☆20Updated last year
- ☆32Updated last year
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆92Updated last year
- Hugging Face Inference Toolkit used to serve transformers, sentence-transformers, and diffusers models.☆51Updated this week
- A library for squeakily cleaning and filtering language datasets.☆45Updated last year
- Minimum Bayes Risk Decoding for Hugging Face Transformers☆56Updated 5 months ago
- Experiments with generating opensource language model assistants☆97Updated last year
- Official implementation of "GPT or BERT: why not both?"☆37Updated last week
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆35Updated 11 months ago
- ☆68Updated 3 months ago
- See https://github.com/cuda-mode/triton-index/ instead!☆11Updated 6 months ago
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆23Updated 8 months ago
- ☆41Updated 2 weeks ago
- PyTorch implementation for MRL☆18Updated 9 months ago