gordicaleksa / Open-NLLBLinks
Effort to open-source NLLB checkpoints.
☆465Updated last year
Alternatives and similar repositories for Open-NLLB
Users that are interested in Open-NLLB are comparing it to the libraries listed below
Sorting:
- ☆552Updated last year
- Repo for the Belebele dataset, a massively multilingual reading comprehension dataset.☆336Updated 11 months ago
- This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and bench…☆598Updated 2 years ago
- LaMini-LM: A Diverse Herd of Distilled Models from Large-Scale Instructions☆824Updated 2 years ago
- ☆1,138Updated 9 months ago
- ☆207Updated last year
- ☆446Updated last year
- Inference code for Mistral and Mixtral hacked up into original Llama implementation☆371Updated last year
- Tune any FALCON in 4-bit☆465Updated 2 years ago
- FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)☆245Updated last year
- Place where folks can contribute to 🤗 community events☆427Updated last year
- ☆158Updated 2 years ago
- FRP Fork☆177Updated 7 months ago
- ☆175Updated last year
- Code for fine-tuning Platypus fam LLMs using LoRA☆630Updated last year
- ☆715Updated last year
- ☆265Updated last year
- Inference code for the paper "Spirit-LM Interleaved Spoken and Written Language Model".☆926Updated last year
- Whisper with Medusa heads☆865Updated 3 months ago
- TTS with The Massively Multilingual Speech (MMS) project☆231Updated last year
- Fine-tune mistral-7B on 3090s, a100s, h100s☆715Updated 2 years ago
- Run inference on MPT-30B using CPU☆576Updated 2 years ago
- Convenience scripts to finetune (chat-)LLaMa3 and other models for any language☆315Updated last year
- ☆415Updated 2 years ago
- A bagel, with everything.☆325Updated last year
- [NeurIPS 22] [AAAI 24] Recurrent Transformer-based long-context architecture.☆775Updated last year
- Fine-tuning LLMs using QLoRA☆266Updated last year
- Finetune ALL LLMs with ALL Adapeters on ALL Platforms!☆331Updated 4 months ago
- An open-source implementation of Google's PaLM models☆818Updated last year
- Extend existing LLMs way beyond the original training length with constant memory usage, without retraining☆732Updated last year