kingabzpro / Deploying-Llama-3.3-70B

Serve Llama 3.3 70B (with AWQ quantization) using vLLM and deploy it on BentoCloud.
21Updated 2 months ago

Alternatives and similar repositories for Deploying-Llama-3.3-70B:

Users that are interested in Deploying-Llama-3.3-70B are comparing it to the libraries listed below