IlyasMoutawwakil / llm-perf-backendLinks
The backend behind the LLM-Perf Leaderboard
☆11Updated last year
Alternatives and similar repositories for llm-perf-backend
Users that are interested in llm-perf-backend are comparing it to the libraries listed below
Sorting:
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.☆32Updated 3 months ago
- experiments with inference on llama☆103Updated last year
- Machine Learning Serving focused on GenAI with simplicity as the top priority.☆59Updated last week
- 🤝 Trade any tensors over the network☆30Updated 2 years ago
- a pipeline for using api calls to agnostically convert unstructured data into structured training data☆32Updated last year
- ML/DL Math and Method notes☆66Updated 2 years ago
- Genalog is an open source, cross-platform python package allowing generation of synthetic document images with custom degradations and te…☆44Updated last year
- 🕹️ Performance Comparison of MLOps Engines, Frameworks, and Languages on Mainstream AI Models.☆139Updated last year
- TitanML Takeoff Server is an optimization, compression and deployment platform that makes state of the art machine learning models access…☆114Updated last year
- Benchmark suite for LLMs from Fireworks.ai☆84Updated last month
- ☆67Updated 9 months ago
- Official repo for the paper PHUDGE: Phi-3 as Scalable Judge. Evaluate your LLMs with or without custom rubric, reference answer, absolute…☆51Updated last year
- ☆198Updated last year
- High level library for batched embeddings generation, blazingly-fast web-based RAG and quantized indexes processing ⚡☆69Updated last month
- Hugging Face Inference Toolkit used to serve transformers, sentence-transformers, and diffusers models.☆89Updated last month
- Check for data drift between two OpenAI multi-turn chat jsonl files.☆39Updated last year
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- Seemless interface of using PyTOrch distributed with Jupyter notebooks☆57Updated 4 months ago
- Fine-tune an LLM to perform batch inference and online serving.☆115Updated 7 months ago
- QLoRA with Enhanced Multi GPU Support☆37Updated 2 years ago
- ☆31Updated last year
- Multi-threaded matrix multiplication and cosine similarity calculations for dense and sparse matrices. Appropriate for calculating the K …☆86Updated last year
- ☆23Updated 2 years ago
- ☆80Updated last year
- ☆53Updated 11 months ago
- Just a bunch of benchmark logs for different LLMs☆119Updated last year
- Optimizing Causal LMs through GRPO with weighted reward functions and automated hyperparameter tuning using Optuna☆59Updated 2 months ago
- Large Language Model Hosting Container☆91Updated 3 months ago
- 💙 Unstructured Data Connectors for Haystack 2.0☆17Updated 2 years ago
- PyTorch implementation for MRL☆20Updated last year