usyd-fsalab / fp6_llmView on GitHub
An efficient GPU support for LLM inference with x-bit quantization (e.g. FP6,FP5).
276Jul 16, 2025Updated 8 months ago

Alternatives and similar repositories for fp6_llm

Users that are interested in fp6_llm are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?