scooter-lacroix / Stan-s-ML-StackLinks
A complete package that provides you with all the components needed to get started of dive deeper into Machine Learning Workloads on Consumer AMD cards, providing CUDA functionality through fully leveraging HIP and ROCm!
☆38Updated last week
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- llama-swap + a minimal ollama compatible api☆26Updated last week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS over OpenAI endpoints.☆211Updated this week
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆211Updated 3 weeks ago
- GPU Power and Performance Manager☆61Updated 11 months ago
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆280Updated this week
- ☆42Updated 2 years ago
- ☆396Updated 6 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆513Updated this week
- A utility that uses Whisper to transcribe videos and various translation APIs to translate the transcribed text and save them as SRT (sub…☆72Updated last year
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆58Updated this week
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated last year
- A tool to determine whether or not your PC can run a given LLM☆164Updated 8 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆29Updated last year
- Web UI for ExLlamaV2☆510Updated 8 months ago
- InferX: Inference as a Service Platform☆136Updated this week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆438Updated this week
- ☆83Updated this week
- NVIDIA Linux open GPU with P2P support☆59Updated 2 weeks ago
- ROCm docker images with fixes/support for legecy architecture gfx803. eg.Radeon RX 590/RX 580/RX 570/RX 480☆76Updated 4 months ago
- Offline tool that processes YouTube videos using WhisperX for automatic transcription and speaker diarization, detects logical fallacies,…☆29Updated last year
- llama.cpp fork with additional SOTA quants and improved performance☆1,246Updated this week
- KoboldCpp Smart Launcher with GPU Layer and Tensor Override Tuning☆28Updated 4 months ago
- Stable Diffusion Docker image preconfigured for usage with AMD Radeon cards☆138Updated last year
- Free and open source pre-trained translation models, including Kurdish, Samoan, Xhosa, Lao, Corsican, Cebuano, Galician, Russian, Belarus…☆87Updated 2 months ago
- A simple tool to anonymize LLM prompts.☆65Updated 8 months ago
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆104Updated 5 months ago
- Model swapping for llama.cpp (or any local OpenAI API compatible server)☆1,655Updated this week
- AMD APU compatible Ollama. Get up and running with OpenAI gpt-oss, DeepSeek-R1, Gemma 3 and other models.☆108Updated this week
- Benchmark for local LLMs with German "Who Wants to Be a Millionaire" questions.☆42Updated last week