lhl / strix-halo-testingLinks
☆138Updated 3 weeks ago
Alternatives and similar repositories for strix-halo-testing
Users that are interested in strix-halo-testing are comparing it to the libraries listed below
Sorting:
- ☆524Updated this week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆451Updated this week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆241Updated last week
- Linux distro for AI computers. Go from bare-metal GPUs to running AI workloads - like vLLM, SGLang, RAG, and Agents - in minutes, fully a…☆315Updated 2 months ago
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆103Updated this week
- llama.cpp fork with additional SOTA quants and improved performance☆1,329Updated this week
- Reliable model swapping for any local OpenAI compatible server - llama.cpp, vllm, etc☆1,899Updated this week
- AI Cluster deployed with Ansible on Random computers with random capabilities☆273Updated 2 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆278Updated 3 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆21Updated this week
- reddacted lets you analyze & sanitize your online footprint using LLMs, PII detection & sentiment analysis to identify anything that migh…☆112Updated 3 months ago
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆97Updated 2 weeks ago
- High-performance lightweight proxy and load balancer for LLM infrastructure. Intelligent routing, automatic failover and unified model di…☆117Updated this week
- ☆226Updated 6 months ago
- GPU Power and Performance Manager☆61Updated last year
- This is a cross-platform desktop application that allows you to chat with locally hosted LLMs and enjoy features like MCP support☆225Updated 3 months ago
- ☆49Updated last month
- Lemonade helps users run local LLMs with the highest performance by configuring state-of-the-art inference engines for their NPUs and GPU…☆1,622Updated this week
- Generate and execute command line commands using LLM☆50Updated 9 months ago
- LLM Client, Server API and UI☆393Updated this week
- Mem0 Integration with OpenWebUI☆46Updated last week
- Docs for GGUF quantization (unofficial)☆312Updated 4 months ago
- ☆257Updated 5 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆571Updated last week
- A persistent local memory for AI, LLMs, or Copilot in VS Code.☆170Updated 3 weeks ago
- A tool to determine whether or not your PC can run a given LLM☆164Updated 9 months ago
- No-code CLI designed for accelerating ONNX workflows☆216Updated 5 months ago
- A web application that converts speech to speech 100% private☆81Updated 5 months ago
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆311Updated 3 months ago
- Interactive, locally hosted tool to migrate Open-WebUI SQLite databases to PostgreSQL☆175Updated last month