fkodom / dilated-attention-pytorch
View external linksLinks

(Unofficial) Implementation of dilated attention from "LongNet: Scaling Transformers to 1,000,000,000 Tokens" (https://arxiv.org/abs/2307.02486)
52Aug 7, 2023Updated 2 years ago

Alternatives and similar repositories for dilated-attention-pytorch

Users that are interested in dilated-attention-pytorch are comparing it to the libraries listed below

Sorting:

Are these results useful?