intel / Enterprise-RAG
Intel® AI for Enterprise RAG converts enterprise data into actionable insights with excellent TCO. Utilizing Intel Gaudi AI accelerators and Intel Xeon processors ensuring streamlined deployment.
☆12Updated last week
Alternatives and similar repositories for Enterprise-RAG:
Users that are interested in Enterprise-RAG are comparing it to the libraries listed below
- Generative AI Examples is a collection of GenAI examples such as ChatQnA, Copilot, which illustrate the pipeline capabilities of the Open…☆342Updated this week
- GenAI components at micro-service level; GenAI service composer to create mega-service☆92Updated this week
- Large Language Model Text Generation Inference on Habana Gaudi☆31Updated this week
- This repo contains documents of the OPEA project☆29Updated this week
- Evaluation, benchmark, and scorecard, targeting for performance on throughput and latency, accuracy on popular evaluation harness, safety…☆25Updated this week
- Intel® Tensor Processing Primitives extension for Pytorch*☆10Updated last week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆166Updated this week
- Containerization and cloud native suite for OPEA☆35Updated this week
- oneAPI Collective Communications Library (oneCCL)☆218Updated last week
- Profiling Tools Interfaces for GPU (PTI for GPU) is a set of Getting Started Documentation and Tools Library to start performance analysi…☆216Updated 2 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆50Updated this week
- OpenAI Triton backend for Intel® GPUs☆157Updated this week
- ☆97Updated this week
- This is a plugin which lets EC2 developers use libfabric as network provider while running NCCL applications.☆160Updated this week
- ☆60Updated last month
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆436Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆58Updated last month
- Reference models for Intel(R) Gaudi(R) AI Accelerator☆159Updated last week
- oneCCL Bindings for Pytorch*☆87Updated 3 weeks ago
- Reference implementations of MLPerf™ inference benchmarks☆1,289Updated this week
- NVIDIA NCCL Tests for Distributed Training☆78Updated this week
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆155Updated this week
- ☆394Updated last week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.☆12Updated last month
- collection of benchmarks to measure basic GPU capabilities☆287Updated 3 weeks ago
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆338Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆57Updated this week
- Run cloud native workloads on NVIDIA GPUs☆153Updated last week
- ☆34Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆201Updated this week