microsoft / deepspeed-gpt-neox
An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library.
☆19Updated 2 years ago
Alternatives and similar repositories for deepspeed-gpt-neox:
Users that are interested in deepspeed-gpt-neox are comparing it to the libraries listed below
- Megatron LM 11B on Huggingface Transformers☆27Updated 3 years ago
- data related codebase for polyglot project☆19Updated last year
- ☆24Updated 2 years ago
- Anh - LAION's multilingual assistant datasets and models☆27Updated last year
- Convenient Text-to-Text Training for Transformers☆19Updated 3 years ago
- ↔️ T5 Machine Translation from English to Korean