openvinotoolkit / openvino_tokenizersLinks
OpenVINO Tokenizers extension
☆46Updated last week
Alternatives and similar repositories for openvino_tokenizers
Users that are interested in openvino_tokenizers are comparing it to the libraries listed below
Sorting:
- Run Generative AI models with simple C++/Python API and using OpenVINO Runtime☆414Updated last week
- A curated list of OpenVINO based AI projects☆178Updated 6 months ago
- 🤗 Optimum Intel: Accelerate inference with Intel optimization tools☆528Updated this week
- This repository contains Dockerfiles, scripts, yaml files, Helm charts, etc. used to scale out AI containers with versions of TensorFlow …☆59Updated last week
- Repository for OpenVINO's extra modules☆161Updated last week
- ONNX Runtime: cross-platform, high performance scoring engine for ML models☆78Updated this week
- No-code CLI designed for accelerating ONNX workflows☆224Updated 7 months ago
- With OpenVINO Test Drive, users can run large language models (LLMs) and models trained by Intel Geti on their devices, including AI PCs …☆35Updated last month
- Large Language Model Text Generation Inference on Habana Gaudi☆34Updated 10 months ago
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtime☆434Updated last month
- The framework to generate a Dockerfile, build, test, and deploy a docker image with OpenVINO™ toolkit.☆68Updated last week
- Intel® Extension for DeepSpeed* is an extension to DeepSpeed that brings feature support with SYCL kernels on Intel GPU(XPU) device. Note…☆63Updated 6 months ago
- Explore our open source AI portfolio! Develop, train, and deploy your AI solutions with performance- and productivity-optimized tools fro…☆62Updated 9 months ago
- Intel® AI Super Builder☆151Updated this week
- Tools for easier OpenVINO development/debugging☆10Updated 6 months ago
- AMD related optimizations for transformer models☆97Updated 3 months ago
- Use safetensors with ONNX 🤗☆82Updated this week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆204Updated last week
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Generative AI extensions for onnxruntime☆930Updated last week
- A scalable inference server for models optimized with OpenVINO™☆816Updated this week
- cudnn_frontend provides a c++ wrapper for the cudnn backend API and samples on how to use it☆673Updated last month
- OpenVINO™ integration with TensorFlow☆178Updated last year
- ☆135Updated this week
- ☆279Updated this week
- AMD's graph optimization engine.☆271Updated this week
- OpenVINO Intel NPU Compiler☆77Updated last week
- oneAPI Specification source files☆210Updated last week
- An innovative library for efficient LLM inference via low-bit quantization☆351Updated last year
- ONNX Script enables developers to naturally author ONNX functions and models using a subset of Python.☆418Updated last week