Training and inference on AWS Trainium and Inferentia chips.
☆263Apr 3, 2026Updated last week
Alternatives and similar repositories for optimum-neuron
Users that are interested in optimum-neuron are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- ☆64Updated this week
- Example code for AWS Neuron SDK developers building inference and training applications☆158Apr 2, 2026Updated last week
- ☆111Jan 16, 2025Updated last year
- ☆13Dec 19, 2025Updated 3 months ago
- Powering AWS purpose-built machine learning chips. Blazing fast and cost effective, natively integrated into PyTorch and TensorFlow and i…☆588Mar 18, 2026Updated 3 weeks ago
- Managed Kubernetes at scale on DigitalOcean • AdDigitalOcean Kubernetes includes the control plane, bandwidth allowance, container registry, automatic updates, and more for free.
- ☆62Updated this week
- ☆39Dec 19, 2024Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆25Mar 5, 2026Updated last month
- ☆17Updated this week
- ☆14Aug 29, 2023Updated 2 years ago
- ☆24Nov 18, 2025Updated 4 months ago
- ☆33Updated this week
- This open-source project delivers a complete pipeline for converting multi-page documents (PDFs/images) into structured JSON using Vision…☆15Updated this week
- Accelerated inference of 🤗 models using FuriosaAI NPU chips.☆27Apr 3, 2026Updated last week
- AI Agents on DigitalOcean Gradient AI Platform • AdBuild production-ready AI agents using customizable tools or access multiple LLMs through a single endpoint. Create custom knowledge bases or connect external data.
- Foundation model benchmarking tool. Run any model on any AWS platform and benchmark for performance across instance type and serving stac…☆255Apr 11, 2025Updated last year
- Large Language Model Hosting Container☆92Updated this week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆209Apr 3, 2026Updated last week
- ☆12Dec 20, 2025Updated 3 months ago
- ☆270Updated this week
- A universal scalable machine learning model deployment solution☆251Updated this week
- 🚀 Accelerate inference and training of 🤗 Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimization…☆3,354Apr 2, 2026Updated last week
- ☆25Apr 7, 2026Updated last week
- Code for Incorporating Relevance Feedback for Information-Seeking Retrieval using Few-Shot Document Re-Ranking, EMNLP 2022, https://aclan…☆14Mar 30, 2026Updated 2 weeks ago
- Deploy open-source AI quickly and easily - Bonus Offer • AdRunpod Hub is built for open source. One-click deployment and autoscaling endpoints without provisioning your own infrastructure.
- Large Language Model Text Generation Inference☆10,830Mar 21, 2026Updated 3 weeks ago
- Google TPU optimizations for transformers models☆136Jan 23, 2026Updated 2 months ago
- ☆19Mar 26, 2026Updated 2 weeks ago
- 🏋️ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of O…☆334Apr 3, 2026Updated last week
- ☆22Mar 27, 2023Updated 3 years ago
- Comprehensive, scalable ML inference architecture using Amazon EKS, leveraging Graviton processors for cost-effective CPU-based inference…☆21Mar 12, 2026Updated last month
- ☆22Apr 7, 2026Updated last week
- Collection of best practices, reference architectures, model training examples and utilities to train large models on AWS.☆407Updated this week
- Github action to connect to tailscale☆20Mar 10, 2026Updated last month
- Serverless GPU API endpoints on Runpod - Bonus Credits • AdSkip the infrastructure headaches. Auto-scaling, pay-as-you-go, no-ops approach lets you focus on innovating your application.
- AWS Neuron Deep Learning Containers (DLCs) are a set of Docker images for training and serving models on AWS Trainium and Inferentia inst…☆21Updated this week
- 🤗 Optimum ONNX: Export your model to ONNX and run inference with ONNX Runtime☆131Apr 2, 2026Updated last week
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Sep 19, 2025Updated 6 months ago
- A lightweight Map for {N}☆13Jun 2, 2024Updated last year
- ☆16Jun 25, 2024Updated last year
- Foundation Model Evaluations Library☆282Aug 7, 2025Updated 8 months ago
- collection of serverless machine learning use cases and examples including Hugging Face transformers, timm, Gradio☆16Dec 16, 2022Updated 3 years ago