opea-project / docs
This repo contains documents of the OPEA project
☆29Updated this week
Alternatives and similar repositories for docs:
Users that are interested in docs are comparing it to the libraries listed below
- Evaluation, benchmark, and scorecard, targeting for performance on throughput and latency, accuracy on popular evaluation harness, safety…☆27Updated this week
- GenAI components at micro-service level; GenAI service composer to create mega-service☆106Updated this week
- Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open…☆368Updated this week
- Large Language Model Text Generation Inference on Habana Gaudi☆31Updated last week
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆157Updated last week
- IBM development fork of https://github.com/huggingface/text-generation-inference☆59Updated 2 months ago
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆191Updated this week
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆112Updated last week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆171Updated this week
- Benchmark suite for LLMs from Fireworks.ai☆66Updated last week
- ☆159Updated this week
- Self-host LLMs with vLLM and BentoML☆87Updated this week
- ☆52Updated 5 months ago
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆146Updated 3 weeks ago
- ☆69Updated 10 months ago
- 📡 Deploy AI models and apps to Kubernetes without developing a hernia☆32Updated 8 months ago
- Inference examples☆35Updated last month
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆38Updated 2 months ago
- This project implements a demonstrator agent that compares the Cache-Augmented Generation (CAG) Framework with traditional Retrieval-Augm…☆25Updated last month
- Creating Generative AI Apps which work☆16Updated 7 months ago
- End-to-End LLM Guide☆101Updated 7 months ago
- BigBertha is an architecture design that demonstrates how automated LLMOps (Large Language Models Operations) can be achieved on any Kube…☆27Updated last year
- ☆65Updated 8 months ago
- Runner in charge of collecting metrics from LLM inference endpoints for the Unify Hub☆17Updated last year
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆220Updated this week
- Iterate fast on your RAG pipelines☆22Updated 2 months ago
- 🚀 Use NVIDIA NIMs with Haystack pipelines☆30Updated 5 months ago
- ☆53Updated last month
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆60Updated 2 months ago