kasper0406 / stablehlo-coreml
Convert StableHLO models into Apple Core ML format
☆17Updated last week
Alternatives and similar repositories for stablehlo-coreml:
Users that are interested in stablehlo-coreml are comparing it to the libraries listed below
- ☆51Updated 8 months ago
- ☆295Updated last week
- A user-friendly tool chain that enables the seamless execution of ONNX models using JAX as the backend.☆109Updated this week
- Stores documents and resources used by the OpenXLA developer community☆120Updated 8 months ago
- MLIR-based partitioning system☆80Updated this week
- jax-triton contains integrations between JAX and OpenAI Triton☆390Updated 2 weeks ago
- A stand-alone implementation of several NumPy dtype extensions used in machine learning.☆257Updated 3 weeks ago
- High-Performance SGEMM on CUDA devices☆90Updated 3 months ago
- Backward compatible ML compute opset inspired by HLO/MHLO☆466Updated last week
- Fastest kernels written from scratch☆236Updated 3 weeks ago
- Ahead of Time (AOT) Triton Math Library☆57Updated last week
- An experimental CPU backend for Triton☆105Updated 2 weeks ago
- extensible collectives library in triton☆85Updated 3 weeks ago
- C API for MLX☆106Updated this week
- ☆163Updated 10 months ago
- ☆200Updated this week
- Fast low-bit matmul kernels in Triton☆291Updated this week
- Experiment of using Tangent to autodiff triton☆78Updated last year
- Applied AI experiments and examples for PyTorch☆261Updated last month
- This repository contains the experimental PyTorch native float8 training UX☆223Updated 8 months ago
- ☆207Updated 3 months ago
- Home for "How To Scale Your Model", a short blog-style textbook about scaling LLMs on TPUs☆242Updated last week
- Explore training for quantized models☆17Updated 3 months ago
- seqax = sequence modeling + JAX☆154Updated 2 weeks ago
- Unified compiler/runtime for interfacing with PyTorch Dynamo.☆99Updated last month
- ☆78Updated 5 months ago
- Write a fast kernel and run it on Discord. See how you compare against the best!☆40Updated this week
- PyTorch per step fault tolerance (actively under development)☆284Updated this week
- KernelBench: Can LLMs Write GPU Kernels? - Benchmark with Torch -> CUDA problems☆268Updated last week
- Repo for AI Compiler team. The intended purpose of this repo is for implementation of a PJRT device.☆14Updated this week