opea-project / docsLinks
This repo contains documents of the OPEA project
☆42Updated 2 weeks ago
Alternatives and similar repositories for docs
Users that are interested in docs are comparing it to the libraries listed below
Sorting:
- Evaluation, benchmark, and scorecard, targeting for performance on throughput and latency, accuracy on popular evaluation harness, safety…☆36Updated 2 weeks ago
- GenAI components at micro-service level; GenAI service composer to create mega-service☆160Updated this week
- Containerization and cloud native suite for OPEA☆67Updated last week
- Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open…☆634Updated this week
- GenAI Studio is a low code platform to enable users to construct, evaluate, and benchmark GenAI applications. The platform also provide c…☆45Updated last month
- Large Language Model Text Generation Inference on Habana Gaudi☆33Updated 3 months ago
- 📡 Deploy AI models and apps to Kubernetes without developing a hernia☆32Updated last year
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆129Updated last month
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆188Updated this week
- IBM development fork of https://github.com/huggingface/text-generation-inference☆60Updated last month
- Place to hack on UI for InstructLab☆32Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆295Updated this week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆47Updated this week
- Template to quickly start working with the BeeAI Framework in Python.☆17Updated last week
- A repository of Dockerfiles, scripts, yaml files, Helm Charts, etc. used to build and scale the sample AI workflows with python, kubernet…☆11Updated last year
- Runner in charge of collecting metrics from LLM inference endpoints for the Unify Hub☆17Updated last year
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆61Updated last week
- Self-host LLMs with vLLM and BentoML☆123Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆76Updated this week
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆184Updated 3 weeks ago
- ☆46Updated this week
- OpenVINO Tokenizers extension☆36Updated last week
- Inference server benchmarking tool☆74Updated 2 months ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆473Updated last week
- Explore our open source AI portfolio! Develop, train, and deploy your AI solutions with performance- and productivity-optimized tools fro…☆44Updated 3 months ago
- ☆222Updated this week
- AI-powered tools to automate code documentation and optimize developer operations.☆37Updated 3 weeks ago
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆59Updated last year
- This project implements a demonstrator agent that compares the Cache-Augmented Generation (CAG) Framework with traditional Retrieval-Augm…☆32Updated 5 months ago
- Setup and Installation Instructions for Habana binaries, docker image creation☆25Updated last month