EleutherAI / oslo
OSLO: Open Source for Large-scale Optimization
β175Updated last year
Alternatives and similar repositories for oslo:
Users that are interested in oslo are comparing it to the libraries listed below
- OSLO: Open Source framework for Large-scale model Optimizationβ306Updated 2 years ago
- Data processing system for polyglotβ92Updated last year
- Large scale 4D parallelism pre-training for π€ transformers in Mixture of Experts *(still work in progress)*β81Updated last year
- Inference code for LLaMA models in JAXβ114Updated 7 months ago
- evolve llm training instruction, from english instruction to any language.β115Updated last year
- β64Updated 2 years ago
- some common Huggingface transformers in maximal update parametrization (Β΅P)β78Updated 2 years ago
- Large-scale language modeling tutorials with PyTorchβ290Updated 3 years ago
- Experiments with generating opensource language model assistantsβ97Updated last year
- JAX implementation of the Llama 2 modelβ213Updated 11 months ago
- Multipack distributed sampler for fast padding-free training of LLMsβ184Updated 5 months ago
- Easy Language Model Pretraining leveraging Huggingface's Transformers and Datasetsβ127Updated 2 years ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)β184Updated 2 years ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pileβ115Updated last year
- Pipeline for pulling and processing online language model pretraining data from the webβ175Updated last year
- β335Updated 9 months ago
- FriendliAI Model Hubβ89Updated 2 years ago
- β58Updated 2 years ago
- manage histories of LLM applied applicationsβ88Updated last year
- [Google Meet] MLLM Arxiv Casual Talkβ55Updated last year
- Sakura-SOLAR-DPO: Merge, SFT, and DPOβ116Updated last year
- A performance library for machine learning applications.β182Updated last year
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!β111Updated last year
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)β116Updated 3 years ago
- data related codebase for polyglot projectβ19Updated last year
- [ICLR 2023] Guess the Instruction! Flipped Learning Makes Language Models Stronger Zero-Shot Learnersβ113Updated 4 months ago
- π₯€π§π»βπCode and dataset for our EMNLP 2023 paper - "SODA: Million-scale Dialogue Distillation with Social Commonsense Contextualizationβ¦β223Updated last year
- Code for Zero-Shot Tokenizer Transferβ119Updated this week
- β181Updated 3 weeks ago