yandex-research / btard
Code for the paper "Secure Distributed Training at Scale" (ICML 2022)
☆14Updated last week
Alternatives and similar repositories for btard:
Users that are interested in btard are comparing it to the libraries listed below
- "Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implemen…☆54Updated 4 years ago
- "Moshpit SGD: Communication-Efficient Decentralized Training on Heterogeneous Unreliable Devices", official implementation☆29Updated last week
- Compression schema for gradients of activations in backward pass☆44Updated last year
- Code for "Practical Low-Rank Communication Compression in Decentralized Deep Learning"☆16Updated 4 years ago
- Memory-efficient transformer. Work in progress.☆19Updated 2 years ago
- Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727☆146Updated 3 months ago
- ☆201Updated last year
- Python library for argument and configuration management☆53Updated 2 years ago
- Code accompanying the NeurIPS 2020 paper: WoodFisher (Singh & Alistarh, 2020)☆48Updated 3 years ago
- Code release for REPAIR: REnormalizing Permuted Activations for Interpolation Repair☆46Updated last year
- PyTorch implementation of HashedNets☆36Updated last year
- ☆45Updated 4 years ago
- Efficient reference implementations of the static & dynamic M-FAC algorithms (for pruning and optimization)☆16Updated 2 years ago
- (ICLR 2023 Spotlight) MPCFormer: fast, performant, and private transformer inference with MPC☆90Updated last year
- ☆71Updated 5 months ago
- Official implementation of Neurips 2020 "Sparse Weight Activation Training" paper.☆27Updated 3 years ago
- ☆17Updated 8 months ago
- Distributed K-FAC Preconditioner for PyTorch☆85Updated this week
- Latest Weight Averaging (NeurIPS HITY 2022)☆28Updated last year
- ☆74Updated 5 years ago
- Efficient LLM Inference Acceleration using Prompting☆46Updated 3 months ago
- The implementation for MLSys 2023 paper: "Cuttlefish: Low-rank Model Training without All The Tuning"☆43Updated last year
- ☆92Updated 2 years ago
- Spartan is an algorithm for training sparse neural network models. This repository accompanies the paper "Spartan Differentiable Sparsity…☆24Updated 2 years ago
- Sparsity support for PyTorch☆33Updated this week
- Code for the paper: Why Transformers Need Adam: A Hessian Perspective☆48Updated 9 months ago
- ☆49Updated last year
- Code for Sanity-Checking Pruning Methods: Random Tickets can Win the Jackpot☆42Updated 4 years ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆79Updated last year
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)☆116Updated 3 years ago