mlcommons / training_results_v4.0
This repository contains the results and code for the MLPerf™ Training v4.0 benchmark.
☆12Updated 7 months ago
Alternatives and similar repositories for training_results_v4.0:
Users that are interested in training_results_v4.0 are comparing it to the libraries listed below
- Reference models for Intel(R) Gaudi(R) AI Accelerator☆158Updated last week
- This repository contains the results and code for the MLPerf™ Training v3.1 benchmark.☆17Updated this week
- NVIDIA's launch, startup, and logging scripts used by our MLPerf Training and HPC submissions☆24Updated last week
- Dolomite Engine is a library for pretraining/finetuning LLMs☆27Updated this week
- Benchmarks to capture important workloads.☆29Updated this week
- oneCCL Bindings for Pytorch*☆87Updated 2 weeks ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆47Updated this week
- This repository contains the results and code for the MLPerf™ Training v2.1 benchmark.☆15Updated last year
- Tools to deploy GPU clusters in the Cloud☆30Updated last year
- MLPerf™ logging library☆32Updated last week
- Distributed preprocessing and data loading for language datasets☆39Updated 9 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.☆43Updated 6 months ago
- ☆14Updated last year
- A Python library transfers PyTorch tensors between CPU and NVMe☆102Updated last month
- Intel Gaudi's Megatron DeepSpeed Large Language Models for training☆13Updated last month
- ☆57Updated 7 months ago
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆58Updated last month
- CloudAI Benchmark Framework☆47Updated this week
- Reference implementations of MLPerf™ HPC training benchmarks☆44Updated 7 months ago
- Distributed ML Optimizer☆30Updated 3 years ago
- ☆16Updated 5 years ago
- Fairring (FAIR + Herring) is a plug-in for PyTorch that provides a process group for distributed training that outperforms NCCL at large …☆63Updated 2 years ago
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆165Updated this week
- Automatically insert nvtx ranges to PyTorch models☆17Updated 3 years ago
- ☆114Updated 10 months ago
- A parallel framework for training deep neural networks☆49Updated this week
- ☆18Updated last month
- This repository contains the results and code for the MLPerf™ Training v1.0 benchmark.☆37Updated 10 months ago
- Fabric Manager packaging for Debian☆14Updated 3 years ago
- Large Language Model Text Generation Inference on Habana Gaudi☆29Updated this week