titanml / takeoff-communityLinks
TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models accessible to everyone.
β114Updated last year
Alternatives and similar repositories for takeoff-community
Users that are interested in takeoff-community are comparing it to the libraries listed below
Sorting:
- π Datasets and models for instruction-tuningβ238Updated 2 years ago
- Mistral + Haystack: build RAG pipelines that rock π€β106Updated last year
- β198Updated last year
- Machine Learning Serving focused on GenAI with simplicity as the top priority.β59Updated last month
- Large Language Model (LLM) Inference API and Chatbotβ126Updated last year
- Multi-threaded matrix multiplication and cosine similarity calculations for dense and sparse matrices. Appropriate for calculating the K β¦β83Updated 11 months ago
- Low latency, High Accuracy, Custom Query routers for Humans and Agents. Built by Prithivi Daβ117Updated 8 months ago
- πΉοΈ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.β139Updated last year
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing β‘β68Updated 2 weeks ago
- β48Updated 2 years ago
- π€ Trade any tensors over the networkβ30Updated 2 years ago
- Command Line Interface for Hugging Face Inference Endpointsβ66Updated last year
- Large Language Model Hosting Containerβ90Updated last month
- Hassle-free ML Pipelines on Kubernetesβ39Updated 2 years ago
- β89Updated 2 years ago
- β80Updated last year
- A Lightweight Library for AI Observabilityβ251Updated 9 months ago
- Lite weight wrapper for the independent implementation of SPLADE++ models for search & retrieval pipelines. Models and Library created byβ¦β33Updated last year
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 linesβ196Updated last year
- experiments with inference on llamaβ103Updated last year
- Framework for building and maintaining self-updating prompts for LLMsβ64Updated last year
- β210Updated 5 months ago
- Code for evaluating with Flow-Judge-v0.1 - an open-source, lightweight (3.8B) language model optimized for LLM system evaluations. Crafteβ¦β78Updated last year
- Chunk your text using gpt4o-mini more accuratelyβ44Updated last year
- β74Updated last year
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year
- LangChain chat model abstractions for dynamic failover, load balancing, chaos engineering, and more!β83Updated last year
- Using LlamaIndex with Ray for productionizing LLM applicationsβ71Updated 2 years ago
- Fine-tune an LLM to perform batch inference and online serving.β114Updated 6 months ago
- Python SDK for experimenting, testing, evaluating & monitoring LLM-powered applications - Parea AI (YC S23)β81Updated 9 months ago