philschmid / huggingface-inferentia2-samplesLinks
☆10Updated 11 months ago
Alternatives and similar repositories for huggingface-inferentia2-samples
Users that are interested in huggingface-inferentia2-samples are comparing it to the libraries listed below
Sorting:
- Examples showing use of NGC containers and models withing Amazon SageMaker☆17Updated 2 years ago
- ☆24Updated this week
- Article about deploying machine learning models using grpc, pytorch and asyncio☆28Updated 2 years ago
- Cortex-compatible model server for Python and TensorFlow☆17Updated 2 years ago
- Question Answering Generative AI application with Large Language Models (LLMs) and Amazon OpenSearch Service☆25Updated 6 months ago
- ☆21Updated 2 weeks ago
- Deploy and scale distributed python applications on Amazon EKS using Ray☆14Updated 2 weeks ago
- ☆13Updated 2 years ago
- Deploy and Scale LLM-based applications☆26Updated last year
- This repository contains samples for fine-tuning embedding models using Amazon SageMaker. Embedding models are useful for tasks such as s…☆12Updated 3 months ago
- ☆14Updated last year
- ☆20Updated 4 years ago
- Large Language Model Hosting Container☆89Updated last week
- ☆44Updated 7 months ago
- Sample code for parallelizing across multiple CPU/GPUs on a single machine to speed up deep learning inference☆33Updated 5 years ago
- Amazon SageMaker Managed Spot Training Examples☆51Updated 11 months ago
- This repository is part of a blog post that guides users through creating a visual search application using Amazon SageMaker and Amazon E…☆11Updated last year
- Core Utilities for NVIDIA Merlin☆19Updated 10 months ago
- This repository shows various ways of deploying a vision model (TensorFlow) from 🤗 Transformers.☆30Updated 2 years ago
- ☆18Updated 4 months ago
- A do-framework project to simplify deployment of Kubeflow on Amazon EKS☆21Updated 3 months ago
- ☆57Updated 3 years ago
- Example code for AWS Neuron SDK developers building inference and training applications☆146Updated last week
- ☆28Updated last year
- 3-Pipeline LLMOps Financial advisor. Steaming pipeline deployed on AWS, 24/7 collects, embeds live-data into QdrantDB. Training pipeline …☆23Updated last month
- Simple and easy stable diffusion inference with LightningModule on GPU, CPU and MPS (Possibly all devices supported by Lightning).☆17Updated last year
- Projects completed under LinuxWorld Informatics Ltd. - MLOps Training.☆12Updated 4 years ago
- Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.☆228Updated this week
- Fast model deployment on AWS Lambda☆14Updated last year
- ☆31Updated 2 years ago