InferX: Inference as a Service Platform
☆189Mar 31, 2026Updated last week
Alternatives and similar repositories for inferx
Users that are interested in inferx are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- ☆179Aug 10, 2025Updated 7 months ago
- OllaDeck is a purple technology stack for Generative AI (text modality) cybersecurity. It provides a comprehensive set of tools for both …☆18Sep 21, 2024Updated last year
- Make Qwen3 Think like Gemini 2.5 Pro | Open webui function☆25May 10, 2025Updated 10 months ago
- ☆210Sep 7, 2025Updated 7 months ago
- Reliable model swapping for any local OpenAI/Anthropic compatible server - llama.cpp, vllm, etc☆3,094Updated this week
- Bare Metal GPUs on DigitalOcean Gradient AI • AdPurpose-built for serious AI teams training foundational models, running large-scale inference, and pushing the boundaries of what's possible.
- ☆13Feb 18, 2024Updated 2 years ago
- A pure and fast NumPy implementation of Mamba with cache support.☆18Jun 16, 2024Updated last year
- Vibe Coded Project Management System☆21Apr 19, 2025Updated 11 months ago
- ☆94Mar 28, 2026Updated last week
- ☆1,331Apr 3, 2026Updated last week
- Manifold is an experimental platform for enabling long horizon workflow automation using teams of AI assistants.☆487Apr 3, 2026Updated last week
- ☆20Sep 28, 2024Updated last year
- Deploy Apollo HF space locally☆40Dec 16, 2024Updated last year
- One command brings a complete pre-wired LLM stack with hundreds of services to explore.☆2,773Updated this week
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting with the flexibility to host WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Cloudways by DigitalOcean.
- ☆19Dec 9, 2025Updated 4 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆1,961Updated this week
- Cleanai (https://github.com/willmil11/cleanai) except I'm making it in c now. Fast and clean from the start this time :)☆17Mar 6, 2026Updated last month
- Self-hosted LLM chatbot arena, with yourself as the only judge☆41Feb 6, 2024Updated 2 years ago
- Serving LLMs in the HF-Transformers format via a PyFlask API☆72Sep 10, 2024Updated last year
- ☆22Aug 9, 2024Updated last year
- Mistral7B playing DOOM☆29Mar 27, 2024Updated 2 years ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆166Jul 5, 2025Updated 9 months ago
- Practical and advanced guide to LLMOps. It provides a solid understanding of large language models’ general concepts, deployment techniqu…☆80Aug 16, 2024Updated last year
- NordVPN Special Discount Offer • AdSave on top-rated NordVPN 1 or 2-year plans with secure browsing, privacy protection, and support for for all major platforms.
- A chat UI for Llama.cpp☆16Mar 11, 2026Updated 3 weeks ago
- Minimal web client for chatting and roleplay with AI characters☆26Aug 21, 2025Updated 7 months ago
- Samples of good AI generated CUDA kernels☆102May 30, 2025Updated 10 months ago
- Quickly and securely turn any Linux box into a build and deployment assistant☆25Oct 3, 2024Updated last year
- ik_llama.cpp's Thireus fork with release builds for macOS/Windows/Ubuntu CPU, Vulkan and CUDA☆94Updated this week
- ☆14Dec 6, 2023Updated 2 years ago
- ☆16May 8, 2025Updated 11 months ago
- EpochFS is a versioned cloud file system with git-like branching, transaction support.☆17Mar 11, 2026Updated 3 weeks ago
- 🌳 MCTS-inspired parallel beam search for conversation optimization. Explore multiple dialogue strategies simultaneously, stress-test a…☆35Jan 18, 2026Updated 2 months ago
- Proton VPN Special Offer - Get 70% off • AdSpecial partner offer. Trusted by over 100 million users worldwide. Tested, Approved and Recommended by Experts.
- Tokenflood is a load testing framework for simulating arbitary loads on instruction-tuned LLMs☆44Mar 20, 2026Updated 2 weeks ago
- Writing Tools, Apple's AI-inspired app, enchants Windows, enhancing your pen with AI LLMs. One hotkey press, system-wide, fixes grammar, …☆27Jul 26, 2025Updated 8 months ago
- ☆114Jun 19, 2025Updated 9 months ago
- Give your local LLM a real memory with a lightweight, fully local memory system. 100% offline and under your control.☆70Sep 16, 2025Updated 6 months ago
- DFloat11 [NeurIPS '25]: Lossless Compression of LLMs and DiTs for Efficient GPU Inference☆617Nov 24, 2025Updated 4 months ago
- vLLM performance dashboard☆44Apr 26, 2024Updated last year
- KV cache store for distributed LLM inference☆402Nov 13, 2025Updated 4 months ago