guolinke / fused_opsLinks
☆10Updated 3 years ago
Alternatives and similar repositories for fused_ops
Users that are interested in fused_ops are comparing it to the libraries listed below
Sorting:
- ☆20Updated 4 years ago
- Unofficially Implements https://arxiv.org/abs/2112.05682 to get Linear Memory Cost on Attention for PyTorch☆12Updated 3 years ago
- A logging tool for deep learning.☆63Updated 8 months ago
- [ICML 2020] code for "PowerNorm: Rethinking Batch Normalization in Transformers" https://arxiv.org/abs/2003.07845☆120Updated 4 years ago
- Distributed preprocessing and data loading for language datasets☆39Updated last year
- Torch Distributed Experimental☆117Updated last year
- ☆37Updated 2 years ago
- (ACL-IJCNLP 2021) Convolutions and Self-Attention: Re-interpreting Relative Positions in Pre-trained Language Models.☆21Updated 3 years ago
- ☆16Updated 9 months ago
- Python pdb for multiple processes☆70Updated 6 months ago
- This package implements THOR: Transformer with Stochastic Experts.☆65Updated 4 years ago
- This repository contains the code for the paper in Findings of EMNLP 2021: "EfficientBERT: Progressively Searching Multilayer Perceptron …☆33Updated 2 years ago
- Implementation of the Triangle Multiplicative module, used in Alphafold2 as an efficient way to mix rows or columns of a 2d feature map, …☆39Updated 4 years ago
- Pytorch library for factorized L0-based pruning.☆45Updated 2 years ago
- FairSeq repo with Apollo optimizer☆114Updated 2 years ago
- Contextual Position Encoding but with some custom CUDA Kernels https://arxiv.org/abs/2405.18719☆22Updated last year
- Code for the paper "Query-Key Normalization for Transformers"☆49Updated 4 years ago
- Block Sparse movement pruning☆81Updated 5 years ago
- The official implementation of You Only Compress Once: Towards Effective and Elastic BERT Compression via Exploit-Explore Stochastic Natu…☆48Updated 4 years ago
- [KDD'22] Learned Token Pruning for Transformers☆102Updated 2 years ago
- Source code for "Efficient Training of BERT by Progressively Stacking"☆113Updated 6 years ago
- Large Scale Distributed Model Training strategy with Colossal AI and Lightning AI☆56Updated 2 years ago
- Code for the ICML'20 paper "Improving Transformer Optimization Through Better Initialization"☆89Updated 4 years ago
- PyTorch Examples repo for "ReZero is All You Need: Fast Convergence at Large Depth"☆62Updated last year
- [JMLR'20] NeurIPS 2019 MicroNet Challenge Efficient Language Modeling, Champion☆41Updated 4 years ago
- Standalone Product Key Memory module in Pytorch - for augmenting Transformer models☆87Updated last month
- ☆29Updated 3 years ago
- Using FlexAttention to compute attention with different masking patterns☆47Updated last year
- An implementation of Transformer with Expire-Span, a circuit for learning which memories to retain☆34Updated 5 years ago
- ☆12Updated 2 years ago