1CatAI / 1Cat-vLLMView on GitHub
vLLM fork for Tesla V100 (SM70) with AWQ 4-bit support, CUDA 12.8 build flow, and validated Qwen3.5 27B/35B deployment on multi-GPU V100.
63Mar 12, 2026Updated last week

Alternatives and similar repositories for 1Cat-vLLM

Users that are interested in 1Cat-vLLM are comparing it to the libraries listed below

Sorting:

Are these results useful?