ray-project / llmperf-leaderboard
☆460Updated last year
Alternatives and similar repositories for llmperf-leaderboard
Users that are interested in llmperf-leaderboard are comparing it to the libraries listed below
Sorting:
- ☆531Updated 8 months ago
- LLMPerf is a library for validating and benchmarking LLMs☆900Updated 5 months ago
- NexusRaven-13B, a new SOTA Open-Source LLM for function calling. This repo contains everything for reproducing our evaluation on NexusRav…☆314Updated last year
- Comparison of Language Model Inference Engines☆217Updated 4 months ago
- Serving multiple LoRA finetuned LLM as one☆1,060Updated last year
- ☆255Updated last week
- RayLLM - LLMs on Ray (Archived). Read README for more info.☆1,261Updated 2 months ago
- 🏋️ A unified multi-backend utility for benchmarking Transformers, Timm, PEFT, Diffusers and Sentence-Transformers with full support of O…☆300Updated this week
- Automatically evaluate your LLMs in Google Colab☆622Updated last year
- batched loras☆342Updated last year
- ☆412Updated last year
- ☆451Updated last year
- A tool for evaluating LLMs☆419Updated last year
- Tutorial for building LLM router☆202Updated 9 months ago
- Inference code for Mistral and Mixtral hacked up into original Llama implementation☆371Updated last year
- This repository contains code and tooling for the Abacus.AI LLM Context Expansion project. Also included are evaluation scripts and bench…☆587Updated last year
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆291Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆350Updated 8 months ago
- [ACL2023] We introduce LLM-Blender, an innovative ensembling framework to attain consistently superior performance by leveraging the dive…☆942Updated 6 months ago
- Benchmarking suite for popular AI APIs☆85Updated 3 months ago
- [ICML 2024] Break the Sequential Dependency of LLM Inference Using Lookahead Decoding☆1,248Updated 2 months ago
- [ICML 2024] SqueezeLLM: Dense-and-Sparse Quantization☆687Updated 9 months ago
- Fast parallel LLM inference for MLX☆187Updated 10 months ago
- A bagel, with everything.☆320Updated last year
- AgentSearch is a framework for powering search agents and enabling customizable local search.☆483Updated last year
- FP16xINT4 LLM inference kernel that can achieve near-ideal ~4x speedups up to medium batchsizes of 16-32 tokens.☆818Updated 8 months ago
- ☆873Updated 8 months ago
- ☆543Updated 5 months ago
- Code for Husky, an open-source language agent that solves complex, multi-step reasoning tasks. Husky v1 addresses numerical, tabular and …☆342Updated 11 months ago
- ☆199Updated last year