DevSecOpsSamples / eks-gpu-autoscalingLinks
GPU Auto Scaling based on Prometheus custom metric on EKS
☆18Updated 3 years ago
Alternatives and similar repositories for eks-gpu-autoscaling
Users that are interested in eks-gpu-autoscaling are comparing it to the libraries listed below
Sorting:
- ☆56Updated last year
- Toolkit for allowing inference and serving with PyTorch on SageMaker. Dockerfiles used for building SageMaker Pytorch Containers are at h…☆142Updated last year
- ☆60Updated last year
- Serve scikit-learn, XGBoost, TensorFlow, and PyTorch models with AWS Lambda container images support.☆100Updated last year
- BIG: Back In the Game of Creative AI☆27Updated 2 years ago
- Amazon SageMaker Managed Spot Training Examples☆51Updated last year
- Compilation of examples of SageMaker inference options and other features.☆72Updated 3 months ago
- This project shows how to serve an ONNX-optimized image classification model as a web service with FastAPI, Docker, and Kubernetes.☆224Updated 3 years ago
- Sample code for parallelizing across multiple CPU/GPUs on a single machine to speed up deep learning inference☆33Updated 5 years ago
- AWS Setup for setting up streaming for OpenAI endpoint☆22Updated 2 years ago
- Hands-on hub to learn techniques to optimize and serve AI models to production the most optimal way.☆14Updated 5 months ago
- Large Language Model Hosting Container☆91Updated 3 months ago
- Amazon SageMaker Local Mode Examples☆263Updated 9 months ago
- This is a short example showing how to utilize Amazon SageMaker's real time endpoints with OpenAI's open source Whisper model for audio t…☆75Updated 2 years ago
- Deploy llama.cpp compatible Generative AI LLMs on AWS Lambda!☆177Updated last year
- This repository aims to showcase how to finetune a FM model in Amazon EKS cluster using, JupyterHub to provision notebooks and craft both…☆51Updated 7 months ago
- MLOps on Amazon EKS☆119Updated this week
- Infrastructure as code for GPU accelerated managed Kubernetes clusters.☆57Updated 9 months ago
- 🚀 End-to-end examples and analysis of deploying LLMs serverless using Modal, Runpod, and Beam☆28Updated last year
- How to use stable diffusion model on AWS Sagemaker☆38Updated 2 years ago
- A helper library to connect into Amazon SageMaker with AWS Systems Manager and SSH (Secure Shell)☆258Updated 6 months ago
- This project simplifies personalized Gen-AI SaaS apps. We fine-tune pre-trained models for users, use single GPUs, and ensure real-time r…☆22Updated 2 years ago
- MLOps example using Amazon SageMaker Pipeline and GitHub Actions☆88Updated 7 months ago
- This repository shows various ways of deploying a vision model (TensorFlow) from 🤗 Transformers.☆30Updated 3 years ago
- MLOps on AWS using Amazon SageMaker Pipelines☆32Updated 2 years ago
- ☆44Updated 2 years ago
- Managing your machine learning lifecycle with MLflow and Amazon SageMaker☆154Updated 3 weeks ago
- TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models access…☆114Updated 2 years ago
- In this course navigates through the LLMOps pipeline, enabling you to preprocess training data for supervised fine-tuning and deploy cust…☆14Updated last year
- ☆35Updated 2 years ago