premAI-io / benchmarksLinks
πΉοΈ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.
β139Updated last year
Alternatives and similar repositories for benchmarks
Users that are interested in benchmarks are comparing it to the libraries listed below
Sorting:
- experiments with inference on llamaβ103Updated last year
- β198Updated last year
- The Batched API provides a flexible and efficient way to process multiple requests in a batch, with a primary focus on dynamic batching oβ¦β154Updated 5 months ago
- Fully fine-tune large models like Mistral, Llama-2-13B, or Qwen-14B completely for freeβ232Updated last year
- A high-throughput and memory-efficient inference and serving engine for LLMsβ267Updated 2 weeks ago
- β138Updated 4 months ago
- Efficient vector database for hundred millions of embeddings.β211Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.β32Updated 3 months ago
- Machine Learning Serving focused on GenAI with simplicity as the top priority.β59Updated 2 months ago
- β210Updated 5 months ago
- A set of scripts and notebooks on LLM finetunning and dataset creationβ112Updated last year
- Manage scalable open LLM inference endpoints in Slurm clustersβ278Updated last year
- A Lightweight Library for AI Observabilityβ252Updated 10 months ago
- Let's build better datasets, together!β267Updated last year
- Fine-tune an LLM to perform batch inference and online serving.β115Updated 6 months ago
- Simple UI for debugging correlations of text embeddingsβ305Updated 6 months ago
- EvolKit is an innovative framework designed to automatically enhance the complexity of instructions used for fine-tuning Large Language Mβ¦β245Updated last year
- Toolkit for attaching, training, saving and loading of new heads for transformer modelsβ293Updated 9 months ago
- A stable, fast and easy-to-use inference library with a focus on a sync-to-async APIβ46Updated last year
- TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models accessβ¦β114Updated last year
- Low-Rank adapter extraction for fine-tuned transformers modelsβ180Updated last year
- Lightweight demos for finetuning LLMs. Powered by π€ transformers and open-source datasets.β78Updated last year
- Self-host LLMs with vLLM and BentoMLβ161Updated 3 weeks ago
- Benchmark various LLM Structured Output frameworks: Instructor, Mirascope, Langchain, LlamaIndex, Fructose, Marvin, Outlines, etc on taskβ¦β180Updated last year
- Minimal example scripts of the Hugging Face Trainer, focused on staying under 150 linesβ196Updated last year
- FineTune LLMs in few lines of code (Text2Text, Text2Speech, Speech2Text)β246Updated last year
- Notus is a collection of fine-tuned LLMs using SFT, DPO, SFT+DPO, and/or any other RLHF techniques, while always keeping a data-first appβ¦β170Updated last year
- Comparison of Language Model Inference Enginesβ238Updated last year
- β66Updated 8 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absoluteβ¦β51Updated last year