ferrumox / foxView on GitHub
High-performance LLM inference engine — drop-in replacement for Ollama with faster multi-turn inference, lower TTFT, and higher throughput through prefix caching and continuous batching.
105Mar 25, 2026Updated this week

Alternatives and similar repositories for fox

Users that are interested in fox are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?