intel / intel-ai-assistant-builderLinks
Intel® AI Assistant Builder
☆136Updated this week
Alternatives and similar repositories for intel-ai-assistant-builder
Users that are interested in intel-ai-assistant-builder are comparing it to the libraries listed below
Sorting:
- No-code CLI designed for accelerating ONNX workflows☆221Updated 6 months ago
- MLPerf Client is a benchmark for Windows, Linux and macOS, focusing on client form factors in ML inference scenarios.☆63Updated last month
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆261Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆398Updated this week
- ☆110Updated this week
- OpenVINO Tokenizers extension☆44Updated this week
- A curated list of OpenVINO based AI projects☆173Updated 5 months ago
- Run LLM Agents on Ryzen AI PCs in Minutes☆812Updated this week
- llama.cpp fork used by GPT4All☆55Updated 10 months ago
- The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PC…☆180Updated 3 weeks ago
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆713Updated this week
- Developer kits reference setup scripts for various kinds of Intel platforms and GPUs☆39Updated last week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆518Updated this week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆56Updated this week
- ☆36Updated this week
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆52Updated 10 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- ☆170Updated 3 weeks ago
- An innovative library for efficient LLM inference via low-bit quantization☆351Updated last year
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆641Updated this week
- Intel® NPU Acceleration Library☆702Updated 7 months ago
- LLM inference in C/C++☆103Updated this week
- ☆144Updated 3 weeks ago
- For individual users, watsonx Code Assistant can access a local IBM Granite model☆37Updated 5 months ago
- AMD related optimizations for transformer models☆96Updated 2 months ago
- ☆274Updated this week
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆88Updated this week
- ☆64Updated last year
- High-Performance Text Deduplication Toolkit☆59Updated 3 months ago
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆220Updated 4 months ago