intel / intel-ai-assistant-builderLinks
Intel® AI Assistant Builder
☆67Updated last week
Alternatives and similar repositories for intel-ai-assistant-builder
Users that are interested in intel-ai-assistant-builder are comparing it to the libraries listed below
Sorting:
- Lightweight Inference server for OpenVINO☆180Updated this week
- No-code CLI designed for accelerating ONNX workflows☆192Updated 3 weeks ago
- ☆90Updated 5 months ago
- An extension that lets the AI take the wheel, allowing it to use the mouse and keyboard, recognize UI elements, and prompt itself :3...no…☆120Updated 7 months ago
- LLM inference in C/C++☆77Updated 3 weeks ago
- llama.cpp fork used by GPT4All☆55Updated 3 months ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 8 months ago
- Simple node proxy for llama-server that enables MCP use☆13Updated 3 weeks ago
- TLS & API keys for your LLM APIs☆16Updated 5 months ago
- RetroChat is a powerful command-line interface for interacting with various AI language models. It provides a seamless experience for eng…☆75Updated this week
- automatically quant GGUF models☆181Updated this week
- Phi4 Multimodal Instruct - OpenAI endpoint and Docker Image for self-hosting☆37Updated 3 months ago
- Docker compose to run vLLM on Windows☆81Updated last year
- A fast batching API to serve LLM models☆181Updated last year
- Convert URLs into LLM-friendly markdown chunks☆64Updated 9 months ago
- A simple github actions script to build a llamafile and uploads to huggingface☆14Updated last year
- AI Tensor Engine for ROCm☆201Updated this week
- Local LLM inference & management server with built-in OpenAI API☆31Updated last year
- This small API downloads and exposes access to NeuML's txtai-wikipedia and full wikipedia datasets, taking in a query and returning full …☆93Updated 2 months ago
- Service for testing out the new Qwen2.5 omni model☆51Updated last month
- GroqFlow provides an automated tool flow for compiling machine learning and linear algebra workloads into Groq programs and executing tho…☆109Updated 3 weeks ago
- Modular, open source LLMOps stack that separates concerns: LiteLLM unifies LLM APIs, manages routing and cost controls, and ensures high-…☆102Updated 3 months ago
- cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server a…☆40Updated this week
- General purpose GPU compute framework built on Vulkan to support 1000s of cross vendor graphics cards (AMD, Qualcomm, NVIDIA & friends). …☆48Updated 3 months ago
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆64Updated 2 weeks ago
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆20Updated last week
- Developer kits reference setup scripts for various kinds of Intel platforms and GPUs☆28Updated this week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆65Updated this week
- V.I.S.O.R., my in-development AI-powered voice assistant with integrated memory!☆36Updated last month
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆43Updated 8 months ago