WapaMario63 / GPTQ-for-LLaMa-ROCmView on GitHub
4 bits quantization of LLaMA using GPTQ, ported to HIP for use in AMD GPUs.
32Oct 4, 2023Updated 2 years ago

Alternatives and similar repositories for GPTQ-for-LLaMa-ROCm

Users that are interested in GPTQ-for-LLaMa-ROCm are comparing it to the libraries listed below

Sorting:

Are these results useful?