onnx / turnkeymlLinks
No-code CLI designed for accelerating ONNX workflows
☆208Updated 2 months ago
Alternatives and similar repositories for turnkeyml
Users that are interested in turnkeyml are comparing it to the libraries listed below
Sorting:
- AI Tensor Engine for ROCm☆260Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆323Updated last week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆331Updated this week
- An innovative library for efficient LLM inference via low-bit quantization☆349Updated 11 months ago
- AMD Ryzen™ AI Software includes the tools and runtime libraries for optimizing and deploying AI inference on AMD Ryzen™ AI powered PCs.☆606Updated last week
- Lightweight Inference server for OpenVINO☆202Updated this week
- Use safetensors with ONNX 🤗☆69Updated last month
- High-speed and easy-use LLM serving framework for local deployment☆117Updated 3 weeks ago
- AMD related optimizations for transformer models☆83Updated last week
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆77Updated 2 weeks ago
- [DEPRECATED] Moved to ROCm/rocm-libraries repo☆111Updated this week
- Run LLM Agents on Ryzen AI PCs in Minutes☆529Updated last week
- LLM training in simple, raw C/HIP for AMD GPUs☆51Updated 11 months ago
- Intel® NPU Acceleration Library☆688Updated 4 months ago
- A curated list of OpenVINO based AI projects☆149Updated last month
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆88Updated this week
- ☆123Updated this week
- Advanced Quantization Algorithm for LLMs and VLMs, with support for CPU, Intel GPU, CUDA and HPU. Seamlessly integrated with Torchao, Tra…☆607Updated this week
- Digest AI is a powerful model analysis tool that extracts insights from your models.☆30Updated 2 months ago
- Generative AI extensions for onnxruntime☆806Updated this week
- llama.cpp fork used by GPT4All☆56Updated 6 months ago
- Development repository for the Triton language and compiler☆127Updated this week
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆485Updated last week
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆376Updated this week
- MLPerf Client is a benchmark for Windows and macOS, focusing on client form factors in ML inference scenarios.☆47Updated 3 weeks ago
- OpenVINO Tokenizers extension☆40Updated this week
- Fast and memory-efficient exact attention☆183Updated 2 weeks ago
- OpenAI Triton backend for Intel® GPUs☆205Updated this week
- Repository of model demos using TT-Buda☆62Updated 4 months ago
- VPTQ, A Flexible and Extreme low-bit quantization algorithm☆652Updated 4 months ago