huggingface / tflite-android-transformers
DistilBERT / GPT-2 for on-device inference thanks to TensorFlow Lite with Android demo apps
ā401Updated last year
Alternatives and similar repositories for tflite-android-transformers:
Users that are interested in tflite-android-transformers are comparing it to the libraries listed below
- š² Transformers android examples (Tensorflow Lite & Pytorch Mobile)ā82Updated last year
- Summarization, translation, sentiment-analysis, text-generation and more at blazing speed using a T5 version implemented in ONNX.ā253Updated 2 years ago
- Accelerated NLP pipelines for fast inference on CPU. Built with Transformers and ONNX runtime.ā126Updated 4 years ago
- State of the art faster Transformer with Tensorflow 2.0 ( NLP, Computer Vision, Audio ).ā85Updated last year
- Project tracking of the "Mobile ML Working Group", for the End-to-End TensorFlow Lite tutorials.ā133Updated 2 years ago
- Repository for the paper "Optimal Subarchitecture Extraction for BERT"ā471Updated 2 years ago
- How to create Selfie2Anime from tflite model to Android.ā74Updated 3 years ago
- ā411Updated last year
- An awesome list of TensorFlow Lite models, samples, tutorials, tools and learning resources.ā1,244Updated 2 years ago
- onnxruntime-extensions: A specialized pre- and post- processing library for ONNX Runtimeā352Updated this week
- TFLite Support is a toolkit that helps users to develop ML and deploy TFLite models onto mobile / ioT devices.ā390Updated last week
- ā” boost inference speed of T5 models by 5x & reduce the model size by 3x.ā572Updated last year
- Interactive Neural Machine Translation-lite (INMT-lite) is a framework to train and develop lite versions (.tflite) of models for neural ā¦ā45Updated last year
- EMNLP 2020: "Dialogue Response Ranking Training with Large-Scale Human Feedback Data"ā336Updated 2 months ago
- FastFormers - highly efficient transformer models for NLUā703Updated last year
- This repository contains notebooks that show the usage of TensorFlow Lite for quantizing deep neural networks.ā171Updated 2 years ago
- Fast Inference Solutions for BLOOMā563Updated 3 months ago
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scaleā154Updated last year
- Daquexian's NNAPI Library. ONNX + Android NNAPIā348Updated 4 years ago
- a fast and user-friendly runtime for transformer inference (Bert, Albert, GPT2, Decoders, etc) on CPU and GPU.ā1,502Updated last year
- Examples for using ONNX Runtime for model training.ā324Updated 3 months ago
- Running BERT without Paddingā468Updated 2 years ago
- A search engine for ParlAI's BlenderBot project (and probably other ones as well)ā131Updated 3 years ago
- PyTorch to TensorFlow Lite converterā182Updated 6 months ago
- An efficient implementation of the popular sequence models for text generation, summarization, and translation tasks. https://arxiv.org/pā¦ā429Updated 2 years ago
- A mobile Implementation of llama.cppā299Updated 11 months ago
- Serving PyTorch 1.0 Models as a Web Server in C++ā226Updated 5 years ago
- TensorFlow code and pre-trained models for BERTā114Updated 4 years ago
- TensorFlow Lite models for MIRNet for low-light image enhancement.ā131Updated 4 years ago
- Scripts to train a bidirectional LSTM with knowledge distillation from BERTā158Updated 5 years ago