yaman / fashion-clip-rsLinks
A complete(grpc service and lib) Rust inference with multilingual embedding support. This version leverages the power of Rust for both GRPC services and as a standalone library, providing highly efficient text and image embeddings.
☆39Updated last year
Alternatives and similar repositories for fashion-clip-rs
Users that are interested in fashion-clip-rs are comparing it to the libraries listed below
Sorting:
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆224Updated last month
- ☆140Updated last year
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated 2 years ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated last year
- Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers, Tiktoken and more. Supports BPE, Unigram and…☆40Updated 2 months ago
- Inference engine for GLiNER models, in Rust☆82Updated last month
- Neural search for web-sites, docs, articles - online!☆146Updated 5 months ago
- HSNW module for Redis☆59Updated 5 years ago
- Rust implementation of Surya☆63Updated 10 months ago
- ⚡️Lightning fast in-memory VectorDB written in rust🦀☆28Updated 9 months ago
- ☆157Updated 2 years ago
- implement llava using candle☆15Updated last year
- Modular Rust transformer/LLM library using Candle☆37Updated last year
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆139Updated last year
- ☆39Updated 3 years ago
- A high performance batching router optimises max throughput for text inference workload☆16Updated 2 years ago
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Updated 3 months ago
- OpenAI compatible API for serving LLAMA-2 model☆218Updated 2 years ago
- Tantivy directory implementation backed by object_store☆37Updated last year
- Using modal.com to process FineWeb-edu data☆20Updated 9 months ago
- Optimizing bit-level Jaccard Index and Population Counts for large-scale quantized Vector Search via Harley-Seal CSA and Lookup Tables☆21Updated 7 months ago
- Fast serverless LLM inference, in Rust.☆108Updated 2 months ago
- Demos of some issues with LangChain.☆31Updated 2 years ago
- ☆12Updated last year
- TensorRT-LLM server with Structured Outputs (JSON) built with Rust☆65Updated 8 months ago
- Rust implementation of Huggingface transformers pipelines using onnxruntime backend with bindings to C# and C.☆41Updated 2 years ago
- Contextualized per-token embeddings☆33Updated 7 months ago
- ☆27Updated last year
- A client library in Rust for Nvidia Triton.☆30Updated 2 years ago