onnx / turnkeymlLinks
No-code CLI designed for accelerating ONNX workflows
☆192Updated 2 weeks ago
Alternatives and similar repositories for turnkeyml
Users that are interested in turnkeyml are comparing it to the libraries listed below
Sorting:
- AI Tensor Engine for ROCm☆201Updated this week
- Lightweight Inference server for OpenVINO☆176Updated last week
- The HIP Environment and ROCm Kit - A lightweight open source build system for HIP and ROCm☆126Updated this week
- vLLM: A high-throughput and memory-efficient inference and serving engine for LLMs☆86Updated last week
- AMD related optimizations for transformer models☆77Updated 7 months ago
- ☆119Updated last year
- ☆525Updated 2 weeks ago
- Onboarding documentation source for the AMD Ryzen™ AI Software Platform. The AMD Ryzen™ AI Software Platform enables developers to take…☆64Updated 2 weeks ago
- An innovative library for efficient LLM inference via low-bit quantization☆348Updated 9 months ago
- llama.cpp fork with additional SOTA quants and improved performance☆519Updated this week
- OpenAI Triton backend for Intel® GPUs☆187Updated this week
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆282Updated this week
- ☆157Updated this week
- Digest AI is a powerful model analysis tool that extracts insights from your models.☆24Updated 3 months ago
- ☆106Updated last month
- Run LLM Agents on Ryzen AI PCs in Minutes☆385Updated 2 weeks ago
- Composable Kernel: Performance Portable Programming Model for Machine Learning Tensor Operators☆401Updated this week
- Development repository for the Triton language and compiler☆122Updated this week
- Repository of model demos using TT-Buda☆62Updated 2 months ago
- Local LLM Server with NPU Acceleration☆36Updated this week
- AMD's graph optimization engine.☆220Updated this week
- A general 2-8 bits quantization toolbox with GPTQ/AWQ/HQQ/VPTQ, and export to onnx/onnx-runtime easily.☆172Updated 2 months ago
- hipBLASLt is a library that provides general matrix-matrix operations with a flexible API and extends functionalities beyond a traditiona…☆97Updated this week
- Use safetensors with ONNX 🤗☆61Updated 3 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆263Updated 7 months ago
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆356Updated this week
- High-speed and easy-use LLM serving framework for local deployment☆108Updated 2 months ago
- Model compression for ONNX☆96Updated 6 months ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆467Updated this week
- Generative AI extensions for onnxruntime☆728Updated this week