conanhujinming / text_dedupLinks
High-Performance Text Deduplication Toolkit
☆60Updated 3 months ago
Alternatives and similar repositories for text_dedup
Users that are interested in text_dedup are comparing it to the libraries listed below
Sorting:
- Input your VRAM and RAM and the toolchain will produce a GGUF model tuned to your system within seconds — flexible model sizing and lowes…☆66Updated last week
- Running Microsoft's BitNet via Electron, React & Astro☆48Updated 2 months ago
- Enhancing LLMs with LoRA☆191Updated last month
- Lightweight Llama 3 8B Inference Engine in CUDA C☆53Updated 8 months ago
- Sparse Inferencing for transformer based LLMs☆215Updated 4 months ago
- ☆52Updated 2 months ago
- Load and run Llama from safetensors files in C☆14Updated last year
- Chat WebUI is an easy-to-use user interface for interacting with AI, and it comes with multiple useful built-in tools such as web search …☆47Updated 3 months ago
- Editor with LLM generation tree exploration☆80Updated 10 months ago
- Local Qwen3 LLM inference. One easy-to-understand file of C source with no dependencies.☆148Updated 5 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templates☆200Updated 2 months ago
- instinct.cpp provides ready to use alternatives to OpenAI Assistant API and built-in utilities for developing AI Agent applications (RAG,…☆54Updated last year
- ☆15Updated 8 months ago
- Simple node proxy for llama-server that enables MCP use☆15Updated 7 months ago
- Run multiple resource-heavy Large Models (LM) on the same machine with limited amount of VRAM/other resources by exposing them on differe…☆85Updated last week
- InferX: Inference as a Service Platform☆142Updated this week
- A platform to self-host AI on easy mode☆179Updated this week
- ☆176Updated 4 months ago
- ☆209Updated 3 months ago
- ☆134Updated 7 months ago
- Inference RWKV v7 in pure C.☆42Updated 2 months ago
- llmbasedos — Local-First OS Where Your AI Agents Wake Up and Work☆278Updated 3 months ago
- Analyze Reddit posts☆25Updated 9 months ago
- From-scratch implementation of OpenAI's GPT-OSS model in Python. No Torch, No GPUs.☆106Updated last month
- ggml implementation of embedding models including SentenceTransformer and BGE☆63Updated last year
- Llama.cpp runner/swapper and proxy that emulates LMStudio / Ollama backends☆49Updated 3 months ago
- REAP: Router-weighted Expert Activation Pruning for SMoE compression☆136Updated last week
- Lightweight C inference for Qwen3 GGUF. Multiturn prefix caching & batch processing.☆19Updated 3 months ago
- Open source LLM UI, compatible with all local LLM providers.☆176Updated last year
- Inference of Mamba models in pure C☆194Updated last year