intel / intel-ai-super-builderLinks
Intel® AI Super Builder
☆159Updated this week
Alternatives and similar repositories for intel-ai-super-builder
Users that are interested in intel-ai-super-builder are comparing it to the libraries listed below
Sorting:
- No-code CLI designed for accelerating ONNX workflows☆227Updated 7 months ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆295Updated this week
- MLPerf Client is a benchmark for Windows, Linux and macOS, focusing on client form factors in ML inference scenarios.☆73Updated 2 months ago
- OpenVINO Tokenizers extension☆48Updated this week
- A curated list of OpenVINO based AI projects☆181Updated 7 months ago
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆60Updated 2 weeks ago
- ☆144Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆428Updated this week
- Developer kits reference setup scripts for various kinds of Intel platforms and GPUs☆42Updated this week
- ☆152Updated last month
- ☆183Updated 2 weeks ago
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆92Updated this week
- Intel® SHMEM - Device initiated shared memory based communication library☆32Updated 2 months ago
- Build AI agents for your PC☆916Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆352Updated last year
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆735Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- Generate a llama-quantize command to copy the quantization parameters of any GGUF☆30Updated 2 weeks ago
- For individual users, watsonx Code Assistant can access a local IBM Granite model☆37Updated 7 months ago
- InferX: Inference as a Service Platform☆156Updated this week
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆238Updated last month
- GenAI components at micro-service level; GenAI service composer to create mega-service☆193Updated 3 weeks ago
- ☆91Updated this week
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆769Updated this week
- ☆137Updated this week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆64Updated 7 months ago
- ☆196Updated 3 months ago
- ☆109Updated 5 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- ☆281Updated this week