opea-project / Enterprise-InferenceLinks
Intel® AI for Enterprise Inference optimizes AI inference services on Intel hardware using Kubernetes Orchestration. It automates LLM model deployment for faster inference, resource provisioning, and optimal settings to simplify processes and reduce manual work.
☆23Updated last week
Alternatives and similar repositories for Enterprise-Inference
Users that are interested in Enterprise-Inference are comparing it to the libraries listed below
Sorting:
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆180Updated 5 months ago
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆194Updated last week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆52Updated last week
- ☆168Updated last week
- GenAI Studio is a low code platform to enable users to construct, evaluate, and benchmark GenAI applications. The platform also provide c…☆50Updated last month
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆63Updated last year
- This repo contains documents of the OPEA project☆44Updated last month
- Evaluation, benchmark, and scorecard, targeting for performance on throughput and latency, accuracy on popular evaluation harness, safety…☆37Updated 2 weeks ago
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Data☆43Updated this week
- GenAI components at micro-service level; GenAI service composer to create mega-service☆178Updated last week
- For individual users, watsonx Code Assistant can access a local IBM Granite model☆35Updated 3 months ago
- ☆13Updated last year
- ☆93Updated 2 months ago
- Build Research and Rag agents with Granite on your laptop☆147Updated last week
- Neo4j Extensions and Integrations with Vertex AI and LangChain☆27Updated 6 months ago
- This repository is a combination of llama workflows and agents together which is a powerful concept.☆17Updated last year
- ☆193Updated this week
- GitHub bot to assist with the taxonomy contribution workflow☆17Updated 11 months ago
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆198Updated 3 months ago
- Notebooks to demo the use of Azure AI Python SDK / LangChain with DeepSeek R1 reasoning model in Azure AI Foundry.☆30Updated 8 months ago
- Tutorial to get started with SkyPilot!☆57Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMs☆15Updated last week
- Route LLM requests to the best model for the task at hand.☆109Updated 3 weeks ago
- How to build an ACP compliant agent that uses MCP as well!☆11Updated 5 months ago
- Learn how multimodal AI merges text, image, and audio for smarter models☆26Updated 8 months ago
- Simplifying the definition and execution, scaling and deployment of pipelines on the cloud.☆234Updated 2 years ago
- Explore our open source AI portfolio! Develop, train, and deploy your AI solutions with performance- and productivity-optimized tools fro…☆53Updated 6 months ago
- Machine Learning using oneAPI. Explores Intel Extensions for scikit-learn* and NumPy, SciPy, Pandas powered by oneAPI☆41Updated last year
- Talk to your CSV: how to Visualize Your Data with Langchain and Streamlit☆29Updated 2 years ago
- GroqFlow provides an automated tool flow for compiling machine learning and linear algebra workloads into Groq programs and executing tho…☆112Updated 2 months ago