titanml / takeoff-communityLinks
TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models accessible to everyone.
β114Updated 2 years ago
Alternatives and similar repositories for takeoff-community
Users that are interested in takeoff-community are comparing it to the libraries listed below
Sorting:
- π Datasets and models for instruction-tuningβ238Updated 2 years ago
- Multi-threaded matrix multiplication and cosine similarity calculations for dense and sparse matrices. Appropriate for calculating the K β¦β86Updated last year
- Machine Learning Serving focused on GenAI with simplicity as the top priority.β59Updated 3 weeks ago
- β198Updated last year
- Mistral + Haystack: build RAG pipelines that rock π€β106Updated last year
- π€ Trade any tensors over the networkβ30Updated 2 years ago
- Large Language Model (LLM) Inference API and Chatbotβ128Updated last year
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Daβ119Updated 9 months ago
- Command Line Interface for Hugging Face Inference Endpointsβ65Updated last year
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing β‘β69Updated 2 months ago
- β80Updated last year
- πΉοΈ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.β138Updated last year
- Using LlamaIndex with Ray for productionizing LLM applicationsβ71Updated 2 years ago
- π Unstructured Data Connectors for Haystack 2.0β17Updated 2 years ago
- β48Updated 2 years ago
- Experimental Code for StructuredRAG: JSON Response Formatting with Large Language Modelsβ114Updated 9 months ago
- Python client library for improving your LLM app accuracyβ97Updated 11 months ago
- β185Updated 2 years ago
- Leverage your LangChain trace data for fine tuningβ46Updated last year
- experiments with inference on llamaβ103Updated last year
- β75Updated 2 years ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- A Lightweight Library for AI Observabilityβ255Updated 11 months ago
- Lite weight wrapper for the independent implementation of SPLADE++ models for search & retrieval pipelines. Models and Library created byβ¦β34Updated last year
- β89Updated 2 years ago
- Python SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)β82Updated 11 months ago
- Data extraction with LLM on CPUβ112Updated 2 years ago
- Efficient vector database for hundred millions of embeddings.β211Updated last year
- β29Updated 2 years ago
- LangChain chat model abstractions for dynamic failover, load balancing, chaos engineering, and more!β84Updated 2 years ago