kevaldekivadiya2415 / textembed
TextEmbed is a REST API crafted for high-throughput and low-latency embedding inference. It accommodates a wide variety of embedding models and frameworks, making it ideal for various natural language processing applications.
☆23Updated 8 months ago
Alternatives and similar repositories for textembed
Users that are interested in textembed are comparing it to the libraries listed below
Sorting:
- Deployment a light and full OpenAI API for production with vLLM to support /v1/embeddings with all embeddings models.☆42Updated 10 months ago
- Code for KaLM-Embedding models☆76Updated last month
- ☆62Updated 9 months ago
- Official repository for RAGViz: Diagnose and Visualize Retrieval-Augmented Generation [EMNLP 2024]☆82Updated 3 months ago
- Client Code Examples, Use Cases and Benchmarks for Enterprise h2oGPTe RAG-Based GenAI Platform☆87Updated 3 weeks ago
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆34Updated last week
- Lightweight continuous batching OpenAI compatibility using HuggingFace Transformers include T5 and Whisper.☆22Updated 2 months ago
- Repository for “PlanRAG: A Plan-then-Retrieval Augmented Generation for Generative Large Language Models as Decision Makers”, NAACL24☆137Updated 11 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆131Updated 10 months ago
- ☆53Updated 11 months ago
- Data preparation code for CrystalCoder 7B LLM☆44Updated last year
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing ⚡☆66Updated 6 months ago
- Evaluation of bm42 sparse indexing algorithm☆65Updated 10 months ago
- Simple examples using Argilla tools to build AI☆52Updated 5 months ago
- ☆43Updated 3 months ago
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆55Updated 7 months ago
- Experimental Code for StructuredRAG: JSON Response Formatting with Large Language Models☆105Updated last month
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafte…☆68Updated 6 months ago
- experiments with inference on llama☆104Updated 11 months ago
- Source code of the paper: RetrievalQA: Assessing Adaptive Retrieval-Augmented Generation for Short-form Open-Domain Question Answering [F…☆62Updated 11 months ago
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async API☆45Updated 7 months ago
- ☆75Updated last year
- A simple service that integrates vLLM with Ray Serve for fast and scalable LLM serving.☆65Updated last year
- ☆22Updated last year
- Ready-to-go containerized RAG service. Implemented with text-embedding-inference + Qdrant/LanceDB.☆64Updated 4 months ago
- ☆34Updated last year
- ☆101Updated 8 months ago
- C++ inference wrappers for running blazing fast embedding services on your favourite serverless like AWS Lambda. By Prithivi Da, PRs welc…☆22Updated last year
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆76Updated 6 months ago
- Meta-Chunking: Learning Efficient Text Segmentation via Logical Perception☆157Updated last month