huggingface / optimum-neuronLinks
Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.
☆236Updated this week
Alternatives and similar repositories for optimum-neuron
Users that are interested in optimum-neuron are comparing it to the libraries listed below
Sorting:
- ☆113Updated 7 months ago
- Example code for AWS Neuron SDK developers building inference and training applications☆148Updated this week
- ☆266Updated 4 months ago
- Large Language Model Hosting Container☆90Updated 2 weeks ago
- Toolkit for allowing inference and serving with PyTorch on SageMaker. Dockerfiles used for building SageMaker Pytorch Containers are at h…☆141Updated 10 months ago
- ☆61Updated 2 weeks ago
- ☆62Updated 4 months ago
- Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and i…☆534Updated last week
- ☆199Updated last year
- experiments with inference on llama☆104Updated last year
- ☆44Updated 9 months ago
- ☆73Updated last year
- Hands-on workshop for distributed training and hosting on SageMaker☆146Updated 3 weeks ago
- A helper library to connect into Amazon SageMaker with AWS Systems Manager and SSH (Secure Shell)☆250Updated last month
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆192Updated last week
- A tool to configure, launch and manage your machine learning experiments.☆182Updated this week
- Google TPU optimizations for transformers models☆118Updated 7 months ago
- ☆56Updated last month
- Manage scalable open LLM inference endpoints in Slurm clusters☆270Updated last year
- Collection of best practices, reference architectures, model training examples and utilities to train large models on AWS.☆334Updated last week
- The package used to build the documentation of our Hugging Face repos☆126Updated this week
- 🏋️ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of O…☆309Updated last week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆19Updated 3 weeks ago
- Let's build better datasets, together!☆261Updated 8 months ago
- Amazon SageMaker Managed Spot Training Examples☆51Updated last year
- ☆40Updated 8 months ago
- JetStream is a throughput and memory optimized engine for LLM inference on XLA devices, starting with TPUs (and GPUs in future -- PRs wel…☆369Updated 2 months ago
- batched loras☆345Updated last year
- Blazing fast training of 🤗 Transformers on Graphcore IPUs☆86Updated last year
- NeurIPS Large Language Model Efficiency Challenge: 1 LLM + 1GPU + 1Day☆256Updated last year