pkusc / zaychik-power-controllerLinks
The Zaychik Power Controller server
☆13Updated last year
Alternatives and similar repositories for zaychik-power-controller
Users that are interested in zaychik-power-controller are comparing it to the libraries listed below
Sorting:
- Spack package repository maintained by Student Cluster Competition Team @ Sun Yat-sen University.☆16Updated 5 months ago
- OpenCAEPoro for ASC 2024☆38Updated 2 years ago
- Documentation for HPC course☆160Updated 7 months ago
- Repository for HPCGame 1st Problems.☆70Updated 2 years ago
- An implementation of HPL-AI Mixed-Precision Benchmark based on hpl-2.3☆29Updated 4 years ago
- Wiki fo HPC☆130Updated 6 months ago
- ☆236Updated last month
- ☆288Updated last week
- ☆14Updated 2 weeks ago
- GoPTX: Fine-grained GPU Kernel Fusion by PTX-level Instruction Flow Weaving☆19Updated 6 months ago
- Summary of the Specs of Commonly Used GPUs for Training and Inference of LLM☆75Updated 6 months ago
- A Throughput-Optimized Pipeline Parallel Inference System for Large Language Models☆46Updated last month
- Documentation for YatCPU☆54Updated 2 years ago
- Linpack: configuration, install, optimization☆16Updated 6 years ago
- 🧪 Ultrafast bisulfite☆38Updated last year
- Flash Attention from Scratch on CUDA Ampere☆129Updated 5 months ago
- Here are my personal paper reading notes (including machine learning systems, AI infrastructure, and other interesting stuffs).☆155Updated 2 weeks ago
- A scheduling framework for multitasking over diverse XPUs, including GPUs, NPUs, ASICs, and FPGAs☆157Updated 3 weeks ago
- The dataset and baseline code for ASC23 LLM inference optimization challenge.☆32Updated 2 years ago
- ☆80Updated 3 weeks ago
- ☆34Updated last year
- gLLM: Global Balanced Pipeline Parallelism System for Distributed LLM Serving with Token Throttling☆53Updated last month
- A hybrid partitioner based quantum circuit simulation system on GPU☆48Updated 3 years ago
- Stepwise optimizations of DGEMM on CPU, reaching performance faster than Intel MKL eventually, even under multithreading.☆163Updated 4 years ago
- NEO is a LLM inference engine built to save the GPU memory crisis by CPU offloading☆84Updated 7 months ago
- Artifact from "Hardware Compute Partitioning on NVIDIA GPUs". THIS IS A FORK OF BAKITAS REPO. I AM NOT ONE OF THE AUTHORS OF THE PAPER.☆55Updated 2 months ago
- Multi-Level Triton Runner supporting Python, IR, PTX, and cubin.☆84Updated 2 weeks ago
- Codes & examples for "CUDA - From Correctness to Performance"☆121Updated last year
- REEF is a GPU-accelerated DNN inference serving system that enables instant kernel preemption and biased concurrent execution in GPU sche…☆104Updated 3 years ago
- WaferLLM: Large Language Model Inference at Wafer Scale☆88Updated last month