slashml / amd_inferenceLinks
Docker-based inference engine for AMD GPUs
☆231Updated 8 months ago
Alternatives and similar repositories for amd_inference
Users that are interested in amd_inference are comparing it to the libraries listed below
Sorting:
- Run and explore Llama models locally with minimal dependencies on CPU☆190Updated 8 months ago
- ☆196Updated last month
- A monitoring station for carnivorous flora.☆128Updated last month
- Run larger LLMs with longer contexts on Apple Silicon by using differentiated precision for KV cache quantization. KVSplit enables 8-bit …☆351Updated last month
- Examples and guides for using the VLM Run API☆279Updated 2 weeks ago
- This is a python implementation for stitching images.☆232Updated 8 months ago