philschmid / deepspeed-sagemaker-example
☆20Updated last year
Alternatives and similar repositories for deepspeed-sagemaker-example:
Users that are interested in deepspeed-sagemaker-example are comparing it to the libraries listed below
- Our open source implementation of MiniLMv2 (https://aclanthology.org/2021.findings-acl.188)☆60Updated last year
- Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.☆217Updated this week
- ☆24Updated 9 months ago
- ☆88Updated last year
- ☆64Updated last year
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!☆111Updated last year
- experiments with inference on llama☆104Updated 7 months ago
- AuditNLG: Auditing Generative AI Language Modeling for Trustworthiness☆97Updated last year
- ☆102Updated this week
- ☆50Updated last month
- ☆242Updated 3 months ago
- This project studies the performance and robustness of language models and task-adaptation methods.☆142Updated 7 months ago
- Example code for AWS Neuron SDK developers building inference and training applications☆132Updated this week
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scale☆154Updated last year
- ☆28Updated last year
- ☆97Updated 2 years ago
- A framework for few-shot evaluation of autoregressive language models.☆102Updated last year
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆93Updated last year
- Techniques used to run BLOOM at inference in parallel☆37Updated 2 years ago
- Code for NeurIPS LLM Efficiency Challenge☆54Updated 9 months ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆66Updated 2 months ago
- A diff tool for language models☆42Updated last year
- ☆50Updated this week
- ☆13Updated last year
- ☆55Updated 2 years ago
- minimal pytorch implementation of bm25 (with sparse tensors)☆97Updated 10 months ago
- Open Instruction Generalist is an assistant trained on massive synthetic instructions to perform many millions of tasks☆207Updated last year
- On the Stability of Fine-tuning BERT: Misconceptions, Explanations, and Strong Baselines☆133Updated last year
- ☆73Updated last year