sasha0552 / nvidia-pstatedLinks
A daemon that automatically manages the performance states of NVIDIA GPUs.
☆101Updated last month
Alternatives and similar repositories for nvidia-pstated
Users that are interested in nvidia-pstated are comparing it to the libraries listed below
Sorting:
- GPU Power and Performance Manager☆62Updated last year
- ☆88Updated last week
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆597Updated last week
- KoboldCpp Smart Launcher with GPU Layer and Tensor Override Tuning☆29Updated 7 months ago
- A library and CLI utilities for managing performance states of NVIDIA GPUs.☆31Updated last year
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆261Updated last week
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- ☆228Updated 7 months ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- Stable Diffusion and Flux in pure C/C++☆24Updated last week
- A local AI companion that uses a collection of free, open source AI models in order to create two virtual companions that will follow you…☆237Updated 2 months ago
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆24Updated last year
- SLOP Detector and analyzer based on dictionary for shareGPT JSON and text☆79Updated 2 weeks ago
- LLM Frontend in a single html file☆671Updated last week
- ☆83Updated 9 months ago
- Web UI for ExLlamaV2☆514Updated 10 months ago
- Y'all thought the dead internet theory wasn't real, but HERE IT IS☆208Updated last year
- An OpenAI API compatible API for chat with image input and questions about the images. aka Multimodal.☆266Updated 9 months ago
- Your Trusty Memory-enabled AI Companion - Simple RAG chatbot optimized for local LLMs | 12 Languages Supported | OpenAI API Compatible☆344Updated 9 months ago
- Fresh builds of llama.cpp with AMD ROCm™ 7 acceleration☆139Updated last week
- Code for Papeg.ai☆227Updated 11 months ago
- InferX: Inference as a Service Platform☆143Updated this week
- Croco.Cpp is fork of KoboldCPP infering GGML/GGUF models on CPU/Cuda with KoboldAI's UI. It's powered partly by IK_LLama.cpp, and compati…☆154Updated last week
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆85Updated this week
- An AI assistant beyond the chat box.☆328Updated last year
- Input your VRAM and RAM and the toolchain will produce a GGUF model tuned to your system within seconds — flexible model sizing and lowes…☆68Updated this week
- Collection of tips for using textgen in various ways☆19Updated last year
- llama-swap + a minimal ollama compatible api☆38Updated last week
- A multimodal, function calling powered LLM webui.☆217Updated last year
- ☆51Updated 10 months ago