scooter-lacroix / Stan-s-ML-StackLinks
A complete install script for the complete AMD ROCm ML stack allowing you to get the most out of your hardware with the least amount of stress!
☆31Updated last month
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆22Updated 10 months ago
- llama-swap + a minimal ollama compatible api☆23Updated this week
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆93Updated 2 months ago
- ☆381Updated 4 months ago
- ☆129Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆209Updated 6 months ago
- AMD APU compatible Ollama. Get up and running with Llama 3.3, DeepSeek-R1, Phi-4, Gemma 3, Mistral Small 3.1 and other large language mod…☆80Updated last week
- Lightweight Inference server for OpenVINO☆198Updated this week
- Model swapping for llama.cpp (or any local OpenAPI compatible server)☆1,370Updated this week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆162Updated last year
- A proxy that hosts multiple single-model runners such as LLama.cpp and vLLM☆11Updated 2 months ago
- A tool to determine whether or not your PC can run a given LLM☆164Updated 6 months ago
- GPU Power and Performance Manager☆61Updated 10 months ago
- General Site for the GFX803 ROCm Stuff☆103Updated 3 weeks ago
- A simple tool to anonymize LLM prompts.☆64Updated 7 months ago
- ☆58Updated 10 months ago
- Linux distro for AI computers. Go from bare-metal GPUs to running AI workloads - like vLLM, SGLang, RAG, and Agents - in minutes, fully a…☆240Updated last week
- Input text from speech in any Linux window, the lean, fast and accurate way, using whisper.cpp OFFLINE. Speak with local LLMs via llama.c…☆126Updated last month
- A platform to self-host AI on easy mode☆159Updated 2 weeks ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆73Updated last week
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆31Updated this week
- A lightweight UI for chatting with Ollama models. Streaming responses, conversation history, and multi-model support.☆114Updated 5 months ago
- No-messing-around sh client for llama.cpp's server☆30Updated last year
- Run LLMs on AMD Ryzen™ AI NPUs. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆129Updated last week
- Handy tool to measure the performance and efficiency of LLMs workloads.☆71Updated 4 months ago
- Make PyTorch models at least run on APUs.☆56Updated last year
- InferX is a Inference Function as a Service Platform☆128Updated this week
- ☆221Updated 3 months ago
- Chat with your current directory's files using a local or API LLM.☆395Updated 2 months ago
- LLM Benchmark for Throughput via Ollama (Local LLMs)☆280Updated 2 weeks ago