scooter-lacroix / Stan-s-ML-StackLinks
A complete package that provides you with all the components needed to get started of dive deeper into Machine Learning Workloads on Consumer AMD cards, providing CUDA functionality through fully leveraging HIP and ROCm!
☆44Updated 2 weeks ago
Alternatives and similar repositories for Stan-s-ML-Stack
Users that are interested in Stan-s-ML-Stack are comparing it to the libraries listed below
Sorting:
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆283Updated last week
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆110Updated 2 months ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆217Updated 2 months ago
- General Site for the GFX803 ROCm Stuff☆145Updated 5 months ago
- Run LLMs on AMD Ryzen™ AI NPUs in minutes. Just like Ollama - but purpose-built and deeply optimized for the AMD NPUs.☆665Updated last week
- llama-swap + a minimal ollama compatible api☆45Updated last week
- ☆426Updated 9 months ago
- InferX: Inference as a Service Platform☆151Updated this week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆622Updated this week
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆33Updated last year
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆24Updated last year
- llama.cpp-gfx906☆85Updated 2 weeks ago
- A utility that uses Whisper to transcribe videos and various translation APIs to translate the transcribed text and save them as SRT (sub…☆74Updated last year
- A tool to determine whether or not your PC can run a given LLM☆167Updated last year
- ☆229Updated 8 months ago
- ☆57Updated last year
- llama.cpp fork with additional SOTA quants and improved performance☆1,553Updated this week
- GPU Power and Performance Manager☆66Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- Manifold is an experimental platform for enabling long horizon workflow automation using teams of AI assistants.☆477Updated this week
- Aggregates compute from spare GPU capacity☆189Updated this week
- ☆89Updated last month
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆65Updated 3 months ago
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆345Updated 11 months ago
- Easy to use interface for the Whisper model optimized for all GPUs!☆450Updated 2 weeks ago
- Open source LLM UI, compatible with all local LLM providers.☆177Updated last year
- Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc☆2,260Updated last week
- Web UI for ExLlamaV2☆513Updated 11 months ago
- This project demonstrates a basic chain-of-thought interaction with any LLM (Large Language Model)☆322Updated last year
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆167Updated last week