olealgoritme / gddr6Links
Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.
☆104Updated 5 months ago
Alternatives and similar repositories for gddr6
Users that are interested in gddr6 are comparing it to the libraries listed below
Sorting:
- Core, Junction, and VRAM temperature reader for Linux + GDDR6/GDDR6X GPUs☆54Updated 5 months ago
- NVIDIA Linux open GPU with P2P support☆60Updated last week
- Fast and memory-efficient exact attention☆193Updated last week
- ☆42Updated 2 years ago
- 8-bit CUDA functions for PyTorch Rocm compatible☆41Updated last year
- ☆43Updated 2 weeks ago
- Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs☆126Updated 2 years ago
- A more memory-efficient rewrite of the HF transformers implementation of Llama for use with quantized weights.☆63Updated 2 years ago
- AMD (Radeon GPU) ROCm based setup for popular AI tools on Ubuntu 24.04.1☆212Updated this week
- 8-bit CUDA functions for PyTorch☆63Updated 3 weeks ago
- Prometheus exporter for Linux based GDDR6/GDDR6X VRAM and GPU Core Hot spot temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆23Updated last year
- Comparison of the output quality of quantization methods, using Llama 3, transformers, GGUF, EXL2.☆165Updated last year
- ☆404Updated 6 months ago
- ☆152Updated 3 months ago
- An optimized quantization and inference library for running LLMs locally on modern consumer-class GPUs☆532Updated this week
- ☆17Updated 10 months ago
- Gpu benchmark☆69Updated 8 months ago
- Make PyTorch models at least run on APUs.☆56Updated last year
- AMD related optimizations for transformer models☆90Updated last month
- LLM-powered lossless compression tool☆288Updated last year
- build scripts for ROCm☆186Updated last year
- An efficent implementation of the method proposed in "The Era of 1-bit LLMs"☆154Updated last year
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆73Updated 8 months ago
- QuIP quantization☆59Updated last year
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆450Updated last week
- A daemon that automatically manages the performance states of NVIDIA GPUs.☆96Updated 3 weeks ago
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆90Updated last week
- NVIDIA Linux open GPU with P2P support☆1,260Updated 4 months ago
- ☆60Updated 4 months ago
- Code for the paper "SparseGPT: Massive Language Models Can Be Accurately Pruned in One-Shot" with LLaMA implementation.☆70Updated 2 years ago