Vision-Empower / Kimi-K2-MiniLinks
A miniaturized version of the Kimi-K2 model optimized for deployment on single H100 GPUs.
☆31Updated last week
Alternatives and similar repositories for Kimi-K2-Mini
Users that are interested in Kimi-K2-Mini are comparing it to the libraries listed below
Sorting:
- ☆22Updated 4 months ago
- Clue inspired puzzles for testing LLM deduction abilities☆38Updated 4 months ago
- OpenCoconut implements a latent reasoning paradigm where we generate thoughts before decoding.☆173Updated 6 months ago
- Scaling Data for SWE-agents☆309Updated this week
- Reverse Engineering Gemma 3n: Google's New Edge-Optimized Language Model☆225Updated last month
- EvaByte: Efficient Byte-level Language Models at Scale☆103Updated 3 months ago
- Gödel Agent: A Self-Referential Agent Framework for Recursive Self-Improvement☆115Updated 5 months ago
- Official code repository for Sketch-of-Thought (SoT)☆125Updated 2 months ago
- GRadient-INformed MoE☆264Updated 9 months ago
- ☆105Updated 2 months ago
- Lightweight toolkit package to train and fine-tune 1.58bit Language models☆81Updated 2 months ago
- LLM based agents with proactive interactions, long-term memory, external tool integration, and local deployment capabilities.☆104Updated this week
- Query-agnostic KV cache eviction: 3–4× reduction in memory and 2× decrease in latency (Qwen3/2.5, Gemma3, LLaMA3)☆91Updated 2 weeks ago
- Multi-Granularity LLM Debugger☆86Updated 2 weeks ago
- ☆286Updated last month
- [NeurIPS'24] SelfCodeAlign: Self-Alignment for Code Generation☆308Updated 5 months ago
- ☆79Updated 8 months ago
- ☆45Updated last year
- ☆162Updated 2 months ago
- Building open version of OpenAI o1 via reasoning traces (Groq, ollama, Anthropic, Gemini, OpenAI, Azure supported) Demo: https://hugging…☆181Updated 9 months ago
- CursorCore: Assist Programming through Aligning Anything☆129Updated 5 months ago
- PyTorch implementation of models from the Zamba2 series.☆184Updated 6 months ago
- Easy to use, High Performant Knowledge Distillation for LLMs☆88Updated 2 months ago
- RWKV-7: Surpassing GPT☆94Updated 8 months ago
- ☆94Updated 7 months ago
- ☆88Updated 8 months ago
- ☆88Updated 2 months ago
- ☆159Updated 10 months ago
- An easy-to-understand framework for LLM samplers that rewind and revise generated tokens☆140Updated 5 months ago
- Transplants vocabulary between language models, enabling the creation of draft models for speculative decoding WITHOUT retraining.☆35Updated last week