conanhujinming / text_dedupLinks
High-Performance Text Deduplication Toolkit
☆59Updated 2 months ago
Alternatives and similar repositories for text_dedup
Users that are interested in text_dedup are comparing it to the libraries listed below
Sorting:
- Lightweight Llama 3 8B Inference Engine in CUDA C☆48Updated 7 months ago
- Simple node proxy for llama-server that enables MCP use☆13Updated 5 months ago
- Running Microsoft's BitNet via Electron, React & Astro☆45Updated last month
- A minimalistic C++ Jinja templating engine for LLM chat templates☆193Updated last month
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆81Updated last week
- ☆48Updated 3 weeks ago
- Enhancing LLMs with LoRA☆172Updated last week
- llm client, server and agent☆73Updated this week
- Editor with LLM generation tree exploration☆77Updated 8 months ago
- Inference RWKV v7 in pure C.☆40Updated 3 weeks ago
- Sparse Inferencing for transformer based LLMs☆201Updated 2 months ago
- Load and run Llama from safetensors files in C☆11Updated last year
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆17Updated last month
- ☆13Updated 6 months ago
- Input your VRAM and RAM and the toolchain will produce a GGUF model tuned to your system within seconds — flexible model sizing and lowes…☆62Updated last week
- Inference of Mamba models in pure C☆192Updated last year
- Guaranteed Structured Output from any Language Model via Hierarchical State Machines☆145Updated 3 weeks ago
- Train your own small bitnet model☆75Updated last year
- instinct.cpp provides ready to use alternatives to OpenAI Assistant API and built-in utilities for developing AI Agent applications (RAG,…☆53Updated last year
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆140Updated 3 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆276Updated 2 months ago
- ☆206Updated last month
- 1.58-bit LLaMa model☆83Updated last year
- 1.58 Bit LLM on Apple Silicon using MLX☆225Updated last year
- Welcome to the official repository of SINQ! A novel, fast and high-quality quantization method designed to make any Large Language Model …☆555Updated this week
- LLM Inference on consumer devices☆125Updated 7 months ago
- A python package for serving LLM on OpenAI-compatible API endpoints with prompt caching using MLX.☆97Updated 4 months ago
- Thin wrapper around GGML to make life easier☆40Updated 4 months ago
- ☆104Updated 2 months ago
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools such as web search …☆46Updated 2 months ago