ferrumox / foxView on GitHub
High-performance LLM inference engine — drop-in replacement for Ollama with faster multi-turn inference, lower TTFT, and higher throughput through prefix caching and continuous batching.
124Apr 7, 2026Updated last week

Alternatives and similar repositories for fox

Users that are interested in fox are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.

Sorting:

Are these results useful?