aiha-lab / Attention-Head-Pruning
View external linksLinks

Layer-wise Pruning of Transformer Heads for Efficient Language Modeling
22Feb 22, 2022Updated 3 years ago

Alternatives and similar repositories for Attention-Head-Pruning

Users that are interested in Attention-Head-Pruning are comparing it to the libraries listed below

Sorting:

Are these results useful?