jundaf2 / INT8-Flash-Attention-FMHA-Quantization

156Updated last year

Related projects

Alternatives and complementary repositories for INT8-Flash-Attention-FMHA-Quantization