fkodom / grouped-query-attention-pytorch
View external linksLinks

(Unofficial) PyTorch implementation of grouped-query attention (GQA) from "GQA: Training Generalized Multi-Query Transformer Models from Multi-Head Checkpoints" (https://arxiv.org/pdf/2305.13245.pdf)
189May 9, 2024Updated last year

Alternatives and similar repositories for grouped-query-attention-pytorch

Users that are interested in grouped-query-attention-pytorch are comparing it to the libraries listed below

Sorting:

Are these results useful?