titanml / takeoff-communityLinks
TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models accessible to everyone.
β114Updated last year
Alternatives and similar repositories for takeoff-community
Users that are interested in takeoff-community are comparing it to the libraries listed below
Sorting:
- πΉοΈ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.β139Updated last year
- Machine Learning Serving focused on GenAI with simplicity as the top priority.β59Updated 2 months ago
- π Datasets and models for instruction-tuningβ238Updated 2 years ago
- π€ Trade any tensors over the networkβ30Updated 2 years ago
- β198Updated last year
- β74Updated last year
- Mistral + Haystack: build RAG pipelines that rock π€β106Updated last year
- Large Language Model (LLM) Inference API and Chatbotβ127Updated last year
- Multi-threaded matrix multiplication and cosine similarity calculations for dense and sparse matrices. Appropriate for calculating the K β¦β85Updated 11 months ago
- β48Updated 2 years ago
- Python client library for improving your LLM app accuracyβ97Updated 10 months ago
- Using LlamaIndex with Ray for productionizing LLM applicationsβ71Updated 2 years ago
- LangChain chat model abstractions for dynamic failover, load balancing, chaos engineering, and more!β84Updated last year
- π Unstructured Data Connectors for Haystack 2.0β17Updated 2 years ago
- Command Line Interface for Hugging Face Inference Endpointsβ66Updated last year
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing β‘β68Updated last month
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Daβ119Updated 8 months ago
- β183Updated last year
- experiments with inference on llamaβ103Updated last year
- Leverage your LangChain trace data for fine tuningβ46Updated last year
- Framework for building and maintaining self-updating prompts for LLMsβ65Updated last year
- PanML is a high level generative AI/ML development and analysis library designed for ease of use and fast experimentation.β116Updated 2 years ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- β‘οΈ A fast and flexible PyTorch inference server that runs locally, on any cloud or AI HW.β146Updated last year
- Lite weight wrapper for the independent implementation of SPLADE++ models for search & retrieval pipelines. Models and Library created byβ¦β33Updated last year
- Supervised instruction finetuning for LLM with HF trainer and Deepspeedβ36Updated 2 years ago
- A collection of LLM services you can self host via docker or modal labs to support your applications developmentβ198Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.β32Updated 3 months ago
- β80Updated last year
- Hassle-free ML Pipelines on Kubernetesβ39Updated 2 years ago