yaman / fashion-clip-rsLinks
A complete(grpc service and lib) Rust inference with multilingual embedding support. This version leverages the power of Rust for both GRPC services and as a standalone library, providing highly efficient text and image embeddings.
☆39Updated last year
Alternatives and similar repositories for fashion-clip-rs
Users that are interested in fashion-clip-rs are comparing it to the libraries listed below
Sorting:
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datas…☆225Updated 3 weeks ago
- ☆140Updated last year
- Unofficial python bindings for the rust llm library. 🐍❤️🦀☆76Updated 2 years ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rust☆79Updated 2 years ago
- GPU accelerated client-side embeddings for vector search, RAG etc.☆65Updated 2 years ago
- Inference engine for GLiNER models, in Rust☆83Updated 2 weeks ago
- Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers, Tiktoken and more. Supports BPE, Unigram and…☆40Updated 3 months ago
- ⚡️Lightning fast in-memory VectorDB written in rust🦀☆29Updated 10 months ago
- Neural search for web-sites, docs, articles - online!☆146Updated 5 months ago
- ☆27Updated last year
- CLIP as a service - Embed image and sentences, object recognition, visual reasoning, image classification and reverse image search☆66Updated 6 months ago
- Using modal.com to process FineWeb-edu data☆20Updated 9 months ago
- implement llava using candle☆15Updated last year
- 🚀 End-to-end examples and analysis of deploying LLMs serverless using Modal, Runpod, and Beam☆28Updated last year
- ☆157Updated 2 years ago
- Rust implementation of Surya☆64Updated 10 months ago
- A library for working with GBNF files☆27Updated 2 months ago
- Inference Llama 2 in one file of zero-dependency, zero-unsafe Rust☆39Updated 2 years ago
- Modular Rust transformer/LLM library using Candle☆38Updated last year
- A high performance batching router optimises max throughput for text inference workload☆16Updated 2 years ago
- A client library in Rust for Nvidia Triton.☆30Updated 2 years ago
- GGML implementation of BERT model with Python bindings and quantization.☆58Updated last year
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆138Updated last year
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedback☆114Updated 10 months ago
- Rust implementation of Huggingface transformers pipelines using onnxruntime backend with bindings to C# and C.☆43Updated 2 years ago
- ☆135Updated last year
- Notebooks using the Neural Magic libraries 📓☆39Updated last year
- utilities for loading and running text embeddings with onnx☆45Updated 5 months ago
- Rust crate for some audio utilities☆26Updated 10 months ago
- Optimizing bit-level Jaccard Index and Population Counts for large-scale quantized Vector Search via Harley-Seal CSA and Lookup Tables☆21Updated 8 months ago