tylerachang / goldfish
Goldfish: Monolingual language models for 350 languages.
☆17Updated 8 months ago
Alternatives and similar repositories for goldfish:
Users that are interested in goldfish are comparing it to the libraries listed below
- ☆12Updated 5 months ago
- ☆13Updated 2 weeks ago
- Trully flash implementation of DeBERTa disentangled attention mechanism.☆47Updated this week
- Using short models to classify long texts☆21Updated 2 years ago
- Efficient Language Model Training through Cross-Lingual and Progressive Transfer Learning☆30Updated 2 years ago
- 🕸 GlotCC Dataset and Pipline -- NeurIPS 2024☆18Updated last month
- ☆45Updated 3 months ago
- ☆22Updated 3 months ago
- ☆20Updated 2 years ago
- BPE modification that implements removing of the intermediate tokens during tokenizer training.☆25Updated 5 months ago
- SWIM-IR is a Synthetic Wikipedia-based Multilingual Information Retrieval training set with 28 million query-passage pairs spanning 33 la…☆48Updated last year
- Ranking of fine-tuned HF models as base models.☆35Updated this week
- [EMNLP'23] Official Code for "FOCUS: Effective Embedding Initialization for Monolingual Specialization of Multilingual Models"☆30Updated 6 months ago
- 🚀🤗 A collection of templates for Hugging Face Spaces☆35Updated last year
- Repository containing the SPIN experiments on the DIBT 10k ranked prompts☆24Updated last year
- Are foundation LMs multilingual knowledge bases? (EMNLP 2023)☆19Updated last year
- Plug-and-play Search Interfaces with Pyserini and Hugging Face☆31Updated last year
- Starbucks: Improved Training for 2D Matryoshka Embeddings☆19Updated 3 months ago
- Embedding Recycling for Language models☆38Updated last year
- Pre-train Static Word Embeddings☆59Updated last month
- Effective Unsupervised Domain Adaptation of Neural Rankers by Diversifying Synthetic Query Generation☆14Updated 2 weeks ago
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆26Updated last year
- Code for the paper "Getting the most out of your tokenizer for pre-training and domain adaptation"☆16Updated last year
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found he…☆31Updated last year
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- ☆49Updated 2 months ago
- IntructIR, a novel benchmark specifically designed to evaluate the instruction following ability in information retrieval models. Our foc…☆32Updated 10 months ago
- BLOOM+1: Adapting BLOOM model to support a new unseen language☆71Updated last year
- ☆26Updated 5 months ago
- ☆24Updated last year