philschmid / optimum-static-quantizationLinks
β28Updated 2 years ago
Alternatives and similar repositories for optimum-static-quantization
Users that are interested in optimum-static-quantization are comparing it to the libraries listed below
Sorting:
- TorchServe+Streamlit for easily serving your HuggingFace NER modelsβ33Updated 3 years ago
- π οΈ Tools for Transformers compression using PyTorch Lightning β‘β84Updated 8 months ago
- [WIP] Behold, semantic-search, built over sentence-transformers to make it easy for search engineers to evaluate, optimise and deploy modβ¦β15Updated 2 years ago
- Large Scale BERT Distillationβ33Updated 2 years ago
- NLP Examples using the π€ librariesβ41Updated 4 years ago
- β87Updated 3 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.β93Updated 2 years ago
- Scripts to convert datasets from various sources to Hugging Face Datasets.β57Updated 2 years ago
- Companion Repo for the Vision Language Modelling YouTube series - https://bit.ly/3PsbsC2 - by Prithivi Da. Open to PRs and collaborationsβ14Updated 2 years ago
- Contains Colab Notebooks show cool use-cases of different GCP ML APIs.β10Updated 4 years ago
- KitanaQA: Adversarial training and data augmentation for neural question-answering modelsβ56Updated 2 years ago
- XtremeDistil framework for distilling/compressing massive multilingual neural network models to tiny and efficient models for AI at scaleβ155Updated last year
- BERT Probe: A python package for probing attention based robustness to character and word based adversarial evaluation. Also, with recipeβ¦β18Updated 3 years ago
- State-of-the-art NLP through transformer models in a modular design and consistent APIs.β45Updated 2 years ago
- Comparing PyTorch, JIT and ONNX for inference with Transformersβ19Updated 4 years ago
- Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.β105Updated 3 years ago
- Accelerated NLP pipelines for fast inference on CPU. Built with Transformers and ONNX runtime.β127Updated 4 years ago
- Code for the paper: Saying No is An Art: Contextualized Fallback Responses for Unanswerable Dialogue Queriesβ19Updated 3 years ago
- A deep learning library based on Pytorch focussed on low resource language research and robustnessβ70Updated 3 years ago
- This project shows how to derive the total number of training tokens from a large text dataset from π€ datasets with Apache Beam and Dataβ¦β27Updated 2 years ago
- β19Updated 4 years ago
- Helper scripts and notes that were used while porting various nlp modelsβ45Updated 3 years ago
- A lightweight wrapper for PyTorch that provides a simple declarative API for context switching between devices, distributed modes, mixed-β¦β67Updated 2 years ago
- Our open source implementation of MiniLMv2 (https://aclanthology.org/2021.findings-acl.188)β61Updated 2 years ago
- Build fast gradio demos of fastai learnersβ35Updated 3 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning Pβ¦β34Updated last year
- Presents an optimized Apache Beam pipeline for generating sentence embeddings (runnable on Cloud Dataflow).β20Updated 3 years ago
- Generating Training Data Made Easyβ43Updated 5 years ago
- Using short models to classify long textsβ21Updated 2 years ago
- Lazy Profiler is a simple utility to collect CPU, GPU, RAM and GPU Memory stats while the program is running.β35Updated 4 years ago