hpcaitech / ColossalAI-Benchmark
Performance benchmarking with ColossalAI
☆39Updated 2 years ago
Alternatives and similar repositories for ColossalAI-Benchmark:
Users that are interested in ColossalAI-Benchmark are comparing it to the libraries listed below
- A Python library transfers PyTorch tensors between CPU and NVMe☆113Updated 4 months ago
- Scalable PaLM implementation of PyTorch☆190Updated 2 years ago
- Automated Parallelization System and Infrastructure for Multiple Ecosystems☆79Updated 4 months ago
- Chimera: bidirectional pipeline parallelism for efficiently training large-scale models.☆62Updated 3 weeks ago
- ☆72Updated 3 years ago
- Official repository for LightSeq: Sequence Level Parallelism for Distributed Training of Long Context Transformers☆207Updated 7 months ago
- ☆103Updated 7 months ago
- ☆27Updated 3 years ago
- PyTorch bindings for CUTLASS grouped GEMM.☆118Updated 3 months ago
- (NeurIPS 2022) Automatically finding good model-parallel strategies, especially for complex models and clusters.☆38Updated 2 years ago
- nnScaler: Compiling DNN models for Parallel Training☆106Updated 2 months ago
- ☆91Updated 7 months ago
- ☆81Updated 3 years ago
- ATC23 AE☆45Updated last year
- ☆79Updated 4 months ago
- Odysseus: Playground of LLM Sequence Parallelism☆68Updated 10 months ago
- Sky Computing: Accelerating Geo-distributed Computing in Federated Learning☆90Updated 2 years ago
- PyTorch bindings for CUTLASS grouped GEMM.☆81Updated 5 months ago
- A simple calculation for LLM MFU.☆34Updated last month
- A resilient distributed training framework☆94Updated last year
- This repository contains integer operators on GPUs for PyTorch.☆201Updated last year
- ☆31Updated last year
- An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).☆247Updated 5 months ago
- PipeTransformer: Automated Elastic Pipelining for Distributed Training of Large-scale Models. ICML 2021☆56Updated 3 years ago
- [IJCAI2023] An automated parallel training system that combines the advantages from both data and model parallelism. If you have any inte…☆51Updated last year
- This repository contains the results and code for the MLPerf™ Training v2.0 benchmark.☆28Updated last year
- Zero Bubble Pipeline Parallelism☆381Updated last week
- PyTorch library for cost-effective, fast and easy serving of MoE models.☆165Updated 3 weeks ago
- ☆42Updated 2 years ago
- ☆78Updated last year