AIS-SNU / Optimus-CC
[ASPLOS'23] Optimus-CC: Efficient Large NLP Model Training with 3D Parallelism Aware Communication Compression
☆6Updated 8 months ago
Alternatives and similar repositories for Optimus-CC:
Users that are interested in Optimus-CC are comparing it to the libraries listed below
- RPCNIC: A High-Performance and Reconfigurable PCIe-attached RPC Accelerator [HPCA2025]☆10Updated 4 months ago
- Sharing the codebase and steps for artifact evaluation/reproduction for MICRO 2024 paper☆9Updated 8 months ago
- ☆23Updated 5 months ago
- ☆10Updated last year
- ☆24Updated last year
- ☆14Updated 3 years ago
- A Cycle-level simulator for M2NDP☆27Updated this week
- ☆12Updated 2 years ago
- [HPCA'24] Smart-Infinity: Fast Large Language Model Training using Near-Storage Processing on a Real System☆44Updated last year
- ☆11Updated 4 months ago
- ☆23Updated 2 years ago
- Ok-Topk is a scheme for distributed training with sparse gradients. Ok-Topk integrates a novel sparse allreduce algorithm (less than 6k c…☆25Updated 2 years ago
- Proteus: A High-Throughput Inference-Serving System with Accuracy Scaling☆11Updated last year
- ☆13Updated 11 months ago
- [ACM EuroSys '23] Fast and Efficient Model Serving Using Multi-GPUs with Direct-Host-Access☆56Updated last year
- ☆30Updated last year
- A Fast Graph Update Library for FPGA-based Dynamic Graph Processing☆9Updated 3 years ago
- ☆14Updated 3 years ago
- PIM-DL: Expanding the Applicability of Commodity DRAM-PIMs for Deep Learning via Algorithm-System Co-Optimization☆29Updated last year
- This serves as a repository for reproducibility of the SC21 paper "In-Depth Analyses of Unified Virtual Memory System for GPU Accelerated…☆31Updated last year
- HW/SW co-designed end-host RPC stack☆20Updated 3 years ago
- ☆9Updated 2 years ago
- (elastic) cuckoo hashing☆14Updated 4 years ago
- A PIM instrumentation, compilation, execution, simulation, and evaluation repository for BLIMP-style architectures.☆18Updated 2 years ago
- A minimum demo for PyTorch distributed extension functionality for collectives.☆11Updated 9 months ago
- LIBRA: Enabling Workload-aware Multi-dimensional Network Topology Optimization for Distributed Training of Large AI Models☆10Updated last year
- A source-to-source compiler for optimizing CUDA dynamic parallelism by aggregating launches☆15Updated 5 years ago
- ☆36Updated last year
- SHADE: Enable Fundamental Cacheability for Distributed Deep Learning Training☆32Updated 2 years ago
- Cupcake: A Compression Scheduler for Scalable Communication-Efficient Distributed Training (MLSys '23)☆9Updated last year