mlcommons / mlperf_clientLinks
MLPerf Client is a benchmark for Windows, Linux and macOS, focusing on client form factors in ML inference scenarios.
☆63Updated last month
Alternatives and similar repositories for mlperf_client
Users that are interested in mlperf_client are comparing it to the libraries listed below
Sorting:
- No-code CLI designed for accelerating ONNX workflows☆221Updated 6 months ago
- Intel® AI Assistant Builder☆136Updated this week
- AMD related optimizations for transformer models☆96Updated 2 months ago
- Transformer GPU VRAM estimator☆67Updated last year
- Train, tune, and infer Bamba model☆137Updated 6 months ago
- LLM inference in C/C++☆103Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆94Updated this week
- Machine Learning Agility (MLAgility) benchmark and benchmarking tools☆40Updated 4 months ago
- ☆219Updated 10 months ago
- A minimalistic C++ Jinja templating engine for LLM chat templates☆200Updated 2 months ago
- CPU inference for the DeepSeek family of large language models in C++☆315Updated 2 months ago
- A collection of all available inference solutions for the LLMs☆93Updated 9 months ago
- Large Language Model Text Generation Inference on Habana Gaudi☆34Updated 9 months ago
- LLM training in simple, raw C/HIP for AMD GPUs☆56Updated last year
- Fully Open Language Models with Stellar Performance☆310Updated last month
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆88Updated this week
- python package of rocm-smi-lib☆24Updated last week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆398Updated this week
- This is the documentation repository for SGLang. It is auto-generated from https://github.com/sgl-project/sglang/tree/main/docs.☆92Updated last week
- The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PC…☆180Updated 3 weeks ago
- An innovative library for efficient LLM inference via low-bit quantization☆351Updated last year
- Benchmarking the serving capabilities of vLLM☆58Updated last year
- Self-host LLMs with vLLM and BentoML☆161Updated 3 weeks ago
- SWE-Bench Pro: Can AI Agents Solve Long-Horizon Software Engineering Tasks?☆228Updated last month
- ☆64Updated last year
- ☆66Updated 6 months ago
- Prepare for DeekSeek R1 inference: Benchmark CPU, DRAM, SSD, iGPU, GPU, ... with efficient code.☆73Updated 10 months ago
- LLM Inference on consumer devices☆128Updated 9 months ago
- Route LLM requests to the best model for the task at hand.☆144Updated this week
- Cray-LM unified training and inference stack.☆22Updated 10 months ago