alexisrozhkov / dilated-self-attention
View external linksLinks

Implementation of the dilated self attention as described in "LongNet: Scaling Transformers to 1,000,000,000 Tokens"
13Jul 23, 2023Updated 2 years ago

Alternatives and similar repositories for dilated-self-attention

Users that are interested in dilated-self-attention are comparing it to the libraries listed below

Sorting:

Are these results useful?