aiha-lab / Attention-Head-Pruning

Layer-wise Pruning of Transformer Heads for Efficient Language Modeling
21Updated 2 years ago

Alternatives and similar repositories for Attention-Head-Pruning:

Users that are interested in Attention-Head-Pruning are comparing it to the libraries listed below