ChengZhang-98 / llm-mixed-q
View external linksLinks

Official implementation of EMNLP'23 paper "Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?"
24Oct 25, 2023Updated 2 years ago

Alternatives and similar repositories for llm-mixed-q

Users that are interested in llm-mixed-q are comparing it to the libraries listed below

Sorting:

Are these results useful?