amd / gaiaLinks
Run LLM Agents on Ryzen AI PCs in Minutes
☆649Updated last week
Alternatives and similar repositories for gaia
Users that are interested in gaia are comparing it to the libraries listed below
Sorting:
- Lemonade helps users run local LLMs with the highest performance by configuring state-of-the-art inference engines for their NPUs and GPU…☆1,488Updated this week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆321Updated last week
- No-code CLI designed for accelerating ONNX workflows☆214Updated 4 months ago
- AI PC starter app for doing AI image creation, image stylizing, and chatbot on a PC powered by an Intel® Arc™ GPU.☆630Updated last week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆450Updated last week
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆665Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,258Updated this week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS over OpenAI endpoints.☆213Updated last week
- ☆374Updated last week
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆504Updated this week
- Intel® AI Assistant Builder☆111Updated this week
- ☆477Updated this week
- Docs for GGUF quantization (unofficial)☆286Updated 3 months ago
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆74Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆212Updated this week
- ☆404Updated 6 months ago
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆298Updated 2 months ago
- ☆413Updated last week
- Download models from the Ollama library, without Ollama☆103Updated 11 months ago
- Model swapping for llama.cpp (or any local OpenAI API compatible server)☆1,690Updated this week
- Review/Check GGUF files and estimate the memory usage and maximum tokens per second.☆209Updated 2 months ago
- VS Code extension for LLM-assisted code/text completion☆1,001Updated this week
- LM inference server implementation based on *.cpp.☆281Updated 2 months ago
- Fully Open Language Models with Stellar Performance☆248Updated 2 months ago
- A platform to self-host AI on easy mode☆170Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆359Updated this week
- VSCode AI coding assistant powered by self-hosted llama.cpp endpoint.☆183Updated 8 months ago
- ☆97Updated last month
- MLPerf Client is a benchmark for Windows and macOS, focusing on client form factors in ML inference scenarios.☆51Updated last week
- vLLM for AMD gfx906 GPUs, e.g. Radeon VII / MI50 / MI60☆290Updated 2 weeks ago