olealgoritme / gddr6Links
Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.
☆109Updated 9 months ago
Alternatives and similar repositories for gddr6
Users that are interested in gddr6 are comparing it to the libraries listed below
Sorting:
- ☆49Updated 2 years ago
- build scripts for ROCm☆186Updated 2 years ago
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- NVIDIA Linux open GPU with P2P support☆126Updated last month
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆69Updated 3 months ago
- 8-bit CUDA functions for PyTorch☆70Updated 4 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆622Updated this week
- Make PyTorch models at least run on APUs.☆56Updated 2 years ago
- ☆426Updated 9 months ago
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆24Updated last year
- Fast and memory-efficient exact attention☆213Updated this week
- Simple monkeypatch to boost AMD Navi 3 GPUs☆47Updated 9 months ago
- Benchmark your GPU with ease☆28Updated last month
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆217Updated 2 months ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆64Updated 2 years ago
- Gpu benchmark☆74Updated last year
- AMD related optimizations for transformer models☆97Updated 3 months ago
- ☆51Updated last month
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆110Updated 2 months ago
- ☆237Updated 2 years ago
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆138Updated 2 years ago
- ☆163Updated 7 months ago
- An innovative library for efficient LLM inference via low-bit quantization☆352Updated last year
- Fast inference engine for Transformer models☆55Updated last year
- Train Llama Loras Easily☆31Updated 2 years ago
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆759Updated this week
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆155Updated last year
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆127Updated last year