jerinphilip / slimtLinks
Inference slice of marian for bergamot's tiny11 models. Faster to compile, and wield. Fewer model-archs than bergamot-translator.
☆11Updated 8 months ago
Alternatives and similar repositories for slimt
Users that are interested in slimt are comparing it to the libraries listed below
Sorting:
- Use an appropriate mix of LLMs based on https://nuenki.app/blog research to translate languages better than any one tool.☆26Updated 3 weeks ago
- GGML implementation of BERT model with Python bindings and quantization.☆56Updated last year
- LLaMA2 + Rust☆12Updated last year
- 🤖 AI-powered CLI for file reorganization. Runs fully locally — no data leaves your machine.☆16Updated 2 weeks ago
- cortex.llamacpp is a high-efficiency C++ inference engine for edge computing. It is a dynamic library that can be loaded by any server a…☆42Updated last week
- ☆21Updated 5 months ago
- llama.cpp gguf file parser for javascript☆43Updated 7 months ago
- A converter and basic tester for rwkv onnx☆42Updated last year
- Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers, Tiktoken and more. Supports BPE, Unigram and…☆26Updated 3 months ago
- A fork of llama3.c used to do some R&D on inferencing☆22Updated 6 months ago
- Neural Solr = Solr 9 + Mighty Inference + Node☆17Updated 3 years ago
- Benchmark scripts for comparing different tokenizers and sentence segmenters of German☆12Updated 2 years ago
- ggml implementation of BERT Embedding☆25Updated last year
- A SQLite extension for working with float and binary vectors. Work in progress!☆22Updated 2 years ago
- A Python library for real-time PostgreSQL event-driven cache invalidation.☆22Updated 2 months ago
- Thin wrapper around GGML to make life easier☆36Updated 3 weeks ago
- A highly customizable, full scale web backend for web-rwkv, built on axum with websocket protocol.☆26Updated last year
- Trying to deconstruct RWKV in understandable terms☆14Updated 2 years ago
- Run Gemini Nano locally on chrome☆22Updated last year
- Experiments with BitNet inference on CPU☆54Updated last year
- Like grep but with natural language queries☆50Updated last year
- A high-performance matcher designed to solve LOGICAL and TEXT VARIATIONS problems in word matching, implemented in Rust.☆17Updated this week
- GPU-targeted vendor-agnostic AI library for Windows, and Mistral model implementation.☆58Updated last year
- Hybrid Search (BM25 & Vector) with SQLite☆19Updated 11 months ago
- Go vs Rust AI bot voice conversation☆28Updated 3 months ago
- [ ggml: Tensor library for machine learning ] written in zig.☆15Updated last year
- Efficient BM25 with DuckDB 🦆☆52Updated 6 months ago
- A versatile Go tool for effortlessly generating mock HTTP APIs for all your needs.☆15Updated 8 months ago
- A low latency, fault tolerant API for accessing LLM's written in C++ using llama.cpp.☆10Updated last month
- A C++ implementation of tinyllama inference on CPU.☆10Updated last year