Since OpenAI and friends refuse to give us a max_ctx param in /models, here's the current context window, input token and output token limits for OpenAI (API), Anthropic, Qwen, Deepseek, llama, Phi, Gemini and Mistral
☆67Dec 20, 2025Updated 3 months ago
Alternatives and similar repositories for llm-context-limits
Users that are interested in llm-context-limits are comparing it to the libraries listed below. We may earn a commission when you buy through links labeled 'Ad' on this page.
Sorting:
- Add/remove knowledge from local files/folder.☆25May 15, 2025Updated 11 months ago
- Yet Another (LLM) Web UI, made with Gemini☆12Dec 25, 2024Updated last year
- ☆15Feb 23, 2026Updated last month
- ☆11Feb 20, 2025Updated last year
- An fully autonomous agent that accesses the browser and performs tasks.☆18Apr 25, 2025Updated 11 months ago
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- A collection of the Tracy line of prompts.☆15Feb 22, 2025Updated last year
- ☆24Jan 22, 2025Updated last year
- A meta-framework for self-improving LLMs with transparent reasoning☆38Dec 10, 2025Updated 4 months ago
- Tools for Open-WebUI☆25May 14, 2025Updated 11 months ago
- Quick start for Open WebUI☆181Nov 11, 2025Updated 5 months ago
- Create text chunks which end at natural stopping points without using a tokenizer☆26Nov 26, 2025Updated 4 months ago
- Moondream MCP Server in Python☆44Jul 2, 2025Updated 9 months ago
- A Field-Theoretic Approach to Unbounded Memory in Large Language Models☆20Apr 15, 2025Updated last year
- Lightweight continuous batching OpenAI compatibility using HuggingFace Transformers include T5 and Whisper.☆29Mar 15, 2025Updated last year
- Wordpress hosting with auto-scaling - Free Trial • AdFully Managed hosting for WordPress and WooCommerce businesses that need reliable, auto-scalable performance. Cloudways SafeUpdates now available.
- ☆21Jan 25, 2025Updated last year
- ☆65Nov 23, 2025Updated 4 months ago
- Easy MCP (Model Context Protocol) servers and AI agents, defined as YAML.☆19Dec 9, 2025Updated 4 months ago
- The Fastest Way to Fine-Tune LLMs Locally☆339Dec 18, 2025Updated 3 months ago
- Analyze your Claude Code/Droid/OpenCode and all providers that use Claude Code or Droid token usage and costs from local JSONL files with…☆63Updated this week
- ☆20Nov 5, 2024Updated last year
- ☆16Mar 11, 2025Updated last year
- Connect Claude Code easily to OpenRouter using a local MCP server.☆37Nov 7, 2025Updated 5 months ago
- tools for bitwarden☆11Sep 28, 2024Updated last year
- Managed hosting for WordPress and PHP on Cloudways • AdManaged hosting for WordPress, Magento, Laravel, or PHP apps, on multiple cloud providers. Deploy in minutes on Cloudways by DigitalOcean.
- ☆212Jan 5, 2026Updated 3 months ago
- A systematic reasoning MCP server implementation for Claude Desktop with beam search and thought evaluation.☆29Feb 6, 2025Updated last year
- Open source tool for transcirption and subtitling, alternative to happyscribe.☆34Feb 12, 2025Updated last year
- ☆40Feb 18, 2024Updated 2 years ago
- Glyphs, acting as collaboratively defined symbols linking related concepts, add a layer of multidimensional semantic richness to user-AI …☆57Feb 10, 2025Updated last year
- Modified Beam Search with periodical restart☆12Sep 12, 2024Updated last year
- Local LLM Powered Recursive Search & Smart Knowledge Explorer☆261Oct 18, 2025Updated 5 months ago
- A lightweight LLaMA.cpp HTTP server Docker image based on Alpine Linux.☆34Apr 9, 2026Updated last week
- setting up Mac as much as automatically!☆11Mar 23, 2026Updated 3 weeks ago
- 1-Click AI Models by DigitalOcean Gradient • AdDeploy popular AI models on DigitalOcean Gradient GPU virtual machines with just a single click. Zero configuration with optimized deployments.
- ☆20Nov 26, 2025Updated 4 months ago
- Yet another frontend for LLM, written using .NET and WinUI 3☆11Sep 14, 2025Updated 7 months ago
- RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best …☆10Nov 3, 2023Updated 2 years ago
- A MCP stdio toolpack for local LLMs☆27Apr 6, 2026Updated last week
- Persistent memory for Claude Code — identity, context, and continuity across sessions☆45Updated this week
- ☆10May 2, 2025Updated 11 months ago
- Protocol for Augmented Memory of Project Artifacts (MCP compatible) - extended☆26Jan 24, 2026Updated 2 months ago