scooter-lacroix / Stan-s-ML-StackLinks
A complete install script for the complete AMD ROCm ML stack allowing you to get the most out of your hardware with the least amount of stress!
☆35Updated this week
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated 11 months ago
- llama-swap + a minimal ollama compatible api☆24Updated this week
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆96Updated 2 weeks ago
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆53Updated 4 months ago
- A tool to determine whether or not your PC can run a given LLM☆165Updated 7 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆210Updated 6 months ago
- Lightweight Inference server for OpenVINO☆211Updated this week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆82Updated this week
- ☆391Updated 5 months ago
- A utility that uses Whisper to transcribe videos and various translation APIs to translate the transcribed text and save them as SRT (sub…☆72Updated last year
- Open source LLM UI, compatible with all local LLM providers.☆174Updated 11 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆493Updated this week
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆164Updated last year
- GPU Power and Performance Manager☆61Updated 11 months ago
- A platform to self-host AI on easy mode☆163Updated this week
- ☆83Updated this week
- KoboldCpp Smart Launcher with GPU Layer and Tensor Override Tuning☆27Updated 4 months ago
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆75Updated 3 months ago
- Run LLMs on AMD Ryzen™ AI NPUs. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆194Updated this week
- Easy to use interface for the Whisper model optimized for all GPUs!☆300Updated last month
- Web UI for ExLlamaV2☆513Updated 7 months ago
- InferX is a Inference Function as a Service Platform☆133Updated this week
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆337Updated 6 months ago
- LLM Frontend in a single html file☆644Updated 8 months ago
- A proxy that hosts multiple single-model runners such as LLama.cpp and vLLM☆11Updated 3 months ago
- Benchmark for local LLMs with German "Who Wants to Be a Millionaire" questions.☆38Updated last week
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆28Updated 11 months ago
- Efforts toward giving Qwen 3 Coder 30B A3B proper agentic tool calling capabilities at or near 100% reliability.☆60Updated last month
- NVIDIA Linux open GPU with P2P support☆50Updated 2 weeks ago
- Everything you need to setup on your AMD system for Machine Learning Stuff☆18Updated last month