richardsliu / ray-on-gkeLinks
☆19Updated last year
Alternatives and similar repositories for ray-on-gke
Users that are interested in ray-on-gke are comparing it to the libraries listed below
Sorting:
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆191Updated this week
- This is suite of the hands-on training materials that shows how to scale CV, NLP, time-series forecasting workloads with Ray.☆428Updated last year
- OpenTelemetry Instrumentation for AI Observability☆579Updated this week
- Metrics to evaluate the quality of responses of your Retrieval Augmented Generation (RAG) applications.☆317Updated last month
- User documentation for KServe.☆108Updated 2 weeks ago
- Unified specification for defining and executing ML workflows, making reproducibility, consistency, and governance easier across the ML p…☆94Updated last year
- Easy, fast and very cheap training and inference on AWS Trainium and Inferentia chips.☆239Updated this week
- Fiddler Auditor is a tool to evaluate language models.☆187Updated last year
- ☆294Updated last month
- ☆239Updated this week
- A Lightweight Library for AI Observability☆251Updated 6 months ago
- Examples on how to use LangChain and Ray☆229Updated 2 years ago
- BigBertha is an architecture design that demonstrates how automated LLMOps (Large Language Models Operations) can be achieved on any Kube…☆28Updated last year
- ☆111Updated 7 months ago
- Fast model deployment on AWS Sagemaker☆16Updated last year
- LLMPerf is a library for validating and benchmarking LLMs☆1,000Updated 9 months ago
- ☆87Updated last year
- ☆16Updated last year
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆558Updated this week
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆137Updated last year
- Triton Model Navigator is an inference toolkit designed for optimizing and deploying Deep Learning models with a focus on NVIDIA GPUs.☆209Updated 4 months ago
- ☆467Updated last year
- An inference server for your machine learning models, including support for multiple frameworks, multi-model serving and more☆839Updated 2 weeks ago
- Fine-tune an LLM to perform batch inference and online serving.☆112Updated 3 months ago
- Finetune LLMs on K8s by using Runbooks☆169Updated last year
- Using LlamaIndex with Ray for productionizing LLM applications☆71Updated 2 years ago
- Check for data drift between two OpenAI multi-turn chat jsonl files.☆37Updated last year
- Tune efficiently any LLM model from HuggingFace using distributed training (multiple GPU) and DeepSpeed. Uses Ray AIR to orchestrate the …☆59Updated 2 years ago
- This project shows how to serve an ONNX-optimized image classification model as a web service with FastAPI, Docker, and Kubernetes.☆220Updated 3 years ago
- Compare MLOps Platforms. Breakdowns of SageMaker, VertexAI, AzureML, Dataiku, Databricks, h2o, kubeflow, mlflow...☆394Updated 2 years ago