ChengZhang-98 / llm-mixed-q

Official implementation of EMNLP'23 paper "Revisiting Block-based Quantisation: What is Important for Sub-8-bit LLM Inference?"
19Updated last year

Alternatives and similar repositories for llm-mixed-q:

Users that are interested in llm-mixed-q are comparing it to the libraries listed below