intel / intel-ai-assistant-builderLinks
Intel® AI Assistant Builder
☆140Updated this week
Alternatives and similar repositories for intel-ai-assistant-builder
Users that are interested in intel-ai-assistant-builder are comparing it to the libraries listed below
Sorting:
- No-code CLI designed for accelerating ONNX workflows☆221Updated 7 months ago
- MLPerf Client is a benchmark for Windows, Linux and macOS, focusing on client form factors in ML inference scenarios.☆67Updated last month
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆270Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆410Updated this week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆58Updated this week
- ☆115Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆351Updated last year
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆89Updated 3 weeks ago
- The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PC…☆180Updated last month
- ☆177Updated last month
- ☆36Updated this week
- ☆132Updated 3 weeks ago
- An Awesome list of oneAPI projects☆154Updated 5 months ago
- GenAI components at micro-service level; GenAI service composer to create mega-service☆191Updated last week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆527Updated this week
- ☆147Updated 3 weeks ago
- Intel® SHMEM - Device initiated shared memory based communication library☆32Updated last month
- This project benchmarks 41 open-source large language models across 19 evaluation tasks using the lm-evaluation-harness library.☆85Updated 4 months ago
- ☆66Updated last year
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Generate a llama-quantize command to copy the quantization parameters of any GGUF☆29Updated 5 months ago
- AI Tensor Engine for ROCm☆330Updated this week
- ☆108Updated 4 months ago
- GenAI Studio is a low code platform to enable users to construct, evaluate, and benchmark GenAI applications. The platform also provide c…☆55Updated 3 weeks ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆63Updated 6 months ago
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang☆96Updated this week
- InferX: Inference as a Service Platform☆146Updated this week
- This repo contains documents of the OPEA project☆43Updated 3 weeks ago
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆126Updated last year
- Intel® NPU Acceleration Library☆700Updated 8 months ago