l4rz / running-nvidia-sxm-gpus-in-consumer-pcsLinks
Running SXM2/SXM3/SXM4 NVidia data center GPUs in consumer PCs
☆129Updated 2 years ago
Alternatives and similar repositories for running-nvidia-sxm-gpus-in-consumer-pcs
Users that are interested in running-nvidia-sxm-gpus-in-consumer-pcs are comparing it to the libraries listed below
Sorting:
- A manual for helping using tesla p40 gpu☆138Updated last year
- Linux based GDDR6/GDDR6X VRAM temperature reader for NVIDIA RTX 3000/4000 series GPUs.☆106Updated 7 months ago
- NVIDIA Linux open GPU with P2P support☆1,285Updated 5 months ago
- LLM training in simple, raw C/HIP for AMD GPUs☆54Updated last year
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆73Updated 9 months ago
- ☆46Updated last month
- ☆53Updated last year
- ☆42Updated this week
- AMD related optimizations for transformer models☆96Updated last month
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆580Updated this week
- ☆48Updated 2 years ago
- NVIDIA Linux open GPU with P2P support☆83Updated 3 weeks ago
- Juice Community Version Public Release☆616Updated 6 months ago
- Make PyTorch models at least run on APUs.☆56Updated last year
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆93Updated this week
- ☆166Updated last month
- ☆449Updated 7 months ago
- Repository of model demos using TT-Buda☆63Updated 7 months ago
- ☆417Updated 7 months ago
- No-code CLI designed for accelerating ONNX workflows☆216Updated 5 months ago
- Inference engine for Intel devices. Serve LLMs, VLMs, Whisper, Kokoro-TTS, Embedding and Rerank models over OpenAI endpoints.☆247Updated 3 weeks ago
- FORK of VLLM for AMD MI25/50/60. A high-throughput and memory-efficient inference and serving engine for LLMs☆65Updated 6 months ago
- build scripts for ROCm☆188Updated last year
- Deep Learning Primitives and Mini-Framework for OpenCL☆204Updated last year
- Reverse engineering the rk3588 npu☆99Updated last year
- Inference code for LLaMA models☆42Updated 2 years ago
- Experimental BitNet Implementation☆73Updated this week
- ☆315Updated 3 months ago
- a simple Flash Attention v2 implementation with ROCM (RDNA3 GPU, roc wmma), mainly used for stable diffusion(ComfyUI) in Windows ZLUDA en…☆48Updated last year
- Ultra low overhead NVIDIA GPU telemetry plugin for telegraf with memory temperature readings.☆63Updated last year