opea-project / Enterprise-InferenceLinks
Intel® AI for Enterprise Inference optimizes AI inference services on Intel hardware using Kubernetes Orchestration. It automates LLM model deployment for faster inference, resource provisioning, and optimal settings to simplify processes and reduce manual work.
☆25Updated last month
Alternatives and similar repositories for Enterprise-Inference
Users that are interested in Enterprise-Inference are comparing it to the libraries listed below
Sorting:
- An Operator for deployment and maintenance of NVIDIA NIMs and NeMo microservices in a Kubernetes environment.☆134Updated this week
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆204Updated this week
- ☆17Updated 4 months ago
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆188Updated 6 months ago
- Helm charts for llm-d☆50Updated 3 months ago
- Model Server for Kepler☆29Updated last month
- A toolkit for discovering cluster network topology.☆81Updated this week
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆131Updated last month
- Kubernetes Operator, ansible playbooks, and production scripts for large-scale AIStore deployments on Kubernetes.☆110Updated this week
- ☆57Updated last week
- A tool to detect infrastructure issues on cloud native AI systems☆51Updated last month
- Containerization and cloud native suite for OPEA☆70Updated last month
- ☆69Updated this week
- ☆38Updated this week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆52Updated this week
- Run cloud native workloads on NVIDIA GPUs☆205Updated last month
- Cloud Native Benchmarking of Foundation Models☆44Updated 3 months ago
- GenAI components at micro-service level; GenAI service composer to create mega-service☆184Updated last week
- Examples for building and running LLM services and applications locally with Podman☆183Updated 3 months ago
- The NVIDIA GPU driver container allows the provisioning of the NVIDIA driver through the use of containers.☆139Updated last week
- ☆40Updated last week
- Carbon Limiting Auto Tuning for Kubernetes☆37Updated last year
- 🎉 An awesome & curated list of best LLMOps tools.☆170Updated this week
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Data☆44Updated this week
- ☆42Updated 2 weeks ago
- ☆104Updated last week
- GenAI Studio is a low code platform to enable users to construct, evaluate, and benchmark GenAI applications. The platform also provide c…☆53Updated 2 months ago
- llm-d benchmark scripts and tooling☆33Updated this week
- ☆267Updated this week
- Terraform examples for deploying HPC clusters on OCI☆58Updated 3 weeks ago