yaman / fashion-clip-rsLinks
A complete(grpc service and lib) Rust inference with multilingual embedding support. This version leverages the power of Rust for both GRPC services and as a standalone library, providing highly efficient text and image embeddings.
β39Updated last year
Alternatives and similar repositories for fashion-clip-rs
Users that are interested in fashion-clip-rs are comparing it to the libraries listed below
Sorting:
- β138Updated last year
- Unofficial python bindings for the rust llm library. πβ€οΈπ¦β75Updated 2 years ago
- High-performance MinHash implementation in Rust with Python bindings for efficient similarity estimation and deduplication of large datasβ¦β201Updated last month
- Inference engine for GLiNER models, in Rustβ65Updated last month
- β26Updated 8 months ago
- A single-binary, GPU-accelerated LLM server (HTTP and WebSocket API) written in Rustβ80Updated last year
- Optimizing bit-level Jaccard Index and Population Counts for large-scale quantized Vector Search via Harley-Seal CSA and Lookup Tablesβ20Updated 3 months ago
- Fast and versatile tokenizer for language models, compatible with SentencePiece, Tokenizers, Tiktoken and more. Supports BPE, Unigram andβ¦β31Updated 5 months ago
- GPU accelerated client-side embeddings for vector search, RAG etc.β66Updated last year
- β156Updated 2 years ago
- implement llava using candleβ15Updated last year
- A client library in Rust for Nvidia Triton.β30Updated 2 years ago
- β39Updated 2 years ago
- Neural search for web-sites, docs, articles - online!β138Updated last month
- Modular Rust transformer/LLM library using Candleβ36Updated last year
- π End-to-end examples and analysis of deploying LLMs serverless using Modal, Runpod, and Beamβ28Updated last year
- Using modal.com to process FineWeb-edu dataβ20Updated 4 months ago
- Rust implementation of Suryaβ60Updated 6 months ago
- πΉοΈ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.β137Updated last year
- This repository has code for fine-tuning LLMs with GRPO specifically for Rust Programming using cargo as feedbackβ103Updated 5 months ago
- Proxy server for triton gRPC server that inferences embedding model in Rustβ21Updated last year
- A miniature version of Modalβ20Updated last year
- Rust implementation of Huggingface transformers pipelines using onnxruntime backend with bindings to C# and C.β39Updated 2 years ago
- utilities for loading and running text embeddings with onnxβ44Updated 2 weeks ago
- β130Updated last year
- open source tooling for AI search and understandingβ51Updated 2 years ago
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.β33Updated 3 months ago
- A library for working with GBNF filesβ24Updated 3 weeks ago
- Run AI models anywhere. https://muna.ai/exploreβ63Updated last week
- TensorRT-LLM server with Structured Outputs (JSON) built with Rustβ58Updated 4 months ago