yandex-research / btard
Code for the paper "Secure Distributed Training at Scale" (ICML 2022)
☆15Updated last month
Alternatives and similar repositories for btard:
Users that are interested in btard are comparing it to the libraries listed below
- Compression schema for gradients of activations in backward pass☆44Updated last year
- "Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts" (NeurIPS 2020), original PyTorch implemen…☆54Updated 4 years ago
- "Moshpit SGD: Communication-Efficient Decentralized Training on Heterogeneous Unreliable Devices", official implementation☆29Updated last month
- Towards Understanding Sharpness-Aware Minimization [ICML 2022]☆35Updated 2 years ago
- Practical low-rank gradient compression for distributed optimization: https://arxiv.org/abs/1905.13727☆146Updated 4 months ago
- ☆17Updated 9 months ago
- Code release for REPAIR: REnormalizing Permuted Activations for Interpolation Repair☆47Updated last year
- Memory-efficient transformer. Work in progress.☆19Updated 2 years ago
- Code for "Training Neural Networks with Fixed Sparse Masks" (NeurIPS 2021).☆58Updated 3 years ago
- Latest Weight Averaging (NeurIPS HITY 2022)☆29Updated last year
- Code accompanying the NeurIPS 2020 paper: WoodFisher (Singh & Alistarh, 2020)☆48Updated 4 years ago
- Code for testing DCT plus Sparse (DCTpS) networks☆14Updated 3 years ago
- A fusion of a linear layer and a cross entropy loss, written for pytorch in triton.☆65Updated 7 months ago
- Revisiting Efficient Training Algorithms For Transformer-based Language Models (NeurIPS 2023)☆79Updated last year
- ☆26Updated last year
- Code for "Practical Low-Rank Communication Compression in Decentralized Deep Learning"☆16Updated 4 years ago
- This is unofficial repository for Towards Efficient and Scalable Sharpness-Aware Minimization.☆36Updated 11 months ago
- Experiments from "The Generalization-Stability Tradeoff in Neural Network Pruning": https://arxiv.org/abs/1906.03728.☆14Updated 4 years ago
- Source code of "What can linearized neural networks actually say about generalization?☆20Updated 3 years ago
- ☆24Updated 8 months ago
- Parameter Efficient Transfer Learning with Diff Pruning☆73Updated 4 years ago
- Code for Sanity-Checking Pruning Methods: Random Tickets can Win the Jackpot☆42Updated 4 years ago
- Python library for argument and configuration management☆54Updated 2 years ago
- SGD with large step sizes learns sparse features [ICML 2023]☆32Updated last year
- Training vision models with full-batch gradient descent and regularization☆37Updated 2 years ago
- ☆71Updated 7 months ago
- PyTorch implementation of HashedNets☆36Updated last year
- ☆93Updated 2 years ago
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)☆30Updated 4 months ago
- Model Fusion via Optimal Transport, NeurIPS 2020☆141Updated 2 years ago