opea-project / Enterprise-InferenceLinks
Intel® AI for Enterprise Inference optimizes AI inference services on Intel hardware using Kubernetes Orchestration. It automates LLM model deployment for faster inference, resource provisioning, and optimal settings to simplify processes and reduce manual work.
☆21Updated 3 weeks ago
Alternatives and similar repositories for Enterprise-Inference
Users that are interested in Enterprise-Inference are comparing it to the libraries listed below
Sorting:
- GenAI Studio is a low code platform to enable users to construct, evaluate, and benchmark GenAI applications. The platform also provide c…☆48Updated last week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆13Updated this week
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆190Updated last week
- For individual users, watsonx Code Assistant can access a local IBM Granite model☆34Updated 2 months ago
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆180Updated 4 months ago
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Data☆42Updated this week
- GenAI components at micro-service level; GenAI service composer to create mega-service☆168Updated this week
- Starter template to develop agents for the beeai platform☆17Updated this week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆52Updated last week
- Large Language Model Text Generation Inference on Habana Gaudi☆34Updated 5 months ago
- GroqFlow provides an automated tool flow for compiling machine learning and linear algebra workloads into Groq programs and executing tho…☆112Updated last month
- How to build an ACP compliant agent that uses MCP as well!☆11Updated 3 months ago
- This NVIDIA RAG blueprint serves as a reference solution for a foundational Retrieval Augmented Generation (RAG) pipeline.☆239Updated 3 weeks ago
- This repository is a combination of llama workflows and agents together which is a powerful concept.☆17Updated last year
- This repo contains documents of the OPEA project☆44Updated last week
- ☆79Updated 2 weeks ago
- Open Source AI with Granite and Granite Code☆22Updated last week
- Learn how multimodal AI merges text, image, and audio for smarter models☆25Updated 7 months ago
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆178Updated last month
- ☆142Updated last week
- Evaluation, benchmark, and scorecard, targeting for performance on throughput and latency, accuracy on popular evaluation harness, safety…☆37Updated last week
- Build Research and Rag agents with Granite on your laptop☆140Updated 3 months ago
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆60Updated last year
- See how to augment LLMs with real-time data for dynamic, context-aware apps - Rag + Agents + GraphRAG.☆128Updated last month
- ☆164Updated last week
- Source code for the IBM Granite AI Model Workshop☆33Updated last week
- HyDE based RAG using NVIDIA NIM.☆16Updated last year
- An HTTP service intended as a backend for an LLM that can run arbitrary pieces of Python code.☆65Updated 2 months ago
- ☆13Updated last year
- Explore our open source AI portfolio! Develop, train, and deploy your AI solutions with performance- and productivity-optimized tools fro…☆50Updated 5 months ago