scooter-lacroix / Stan-s-ML-StackLinks
A complete package that provides you with all the components needed to get started of dive deeper into Machine Learning Workloads on Consumer AMD cards, providing CUDA functionality through fully leveraging HIP and ROCm!
☆40Updated last month
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated last year
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆96Updated last month
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆212Updated this week
- llama-swap + a minimal ollama compatible api☆30Updated last week
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆226Updated this week
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆378Updated this week
- ☆42Updated 2 years ago
- ☆409Updated 6 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆81Updated last week
- GPU Power and Performance Manager☆60Updated last year
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆79Updated this week
- General Site for the GFX803 ROCm Stuff☆120Updated 2 months ago
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆55Updated last week
- A tool to determine whether or not your PC can run a given LLM☆164Updated 9 months ago
- ☆84Updated 3 weeks ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- A utility that uses Whisper to transcribe videos and various translation APIs to translate the transcribed text and save them as SRT (sub…☆73Updated last year
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆541Updated last week
- ☆415Updated last week
- Privacy-first agentic framework with powerful reasoning & task automation capabilities. Natively distributed and fully ISO 27XXX complian…☆66Updated 6 months ago
- Reliable model swapping for any local OpenAI compatible server - llama.cpp, vllm, etc☆1,764Updated this week
- llm client, server and agent☆73Updated this week
- InferX: Inference as a Service Platform☆137Updated this week
- Web UI for ExLlamaV2☆511Updated 8 months ago
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- KoboldCpp Smart Launcher with GPU Layer and Tensor Override Tuning☆29Updated 5 months ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆29Updated last year
- Code for Papeg.ai☆225Updated 9 months ago
- German "Who Wants To Be A Millionaire" LLM Benchmarking.☆46Updated last week
- No-messing-around sh client for llama.cpp's server☆30Updated last year