ENOT-AutoDL / gpt-j-6B-tensorrt-int8Links
GPT-J 6B inference on TensorRT with INT-8 precision
☆11Updated 2 years ago
Alternatives and similar repositories for gpt-j-6B-tensorrt-int8
Users that are interested in gpt-j-6B-tensorrt-int8 are comparing it to the libraries listed below
Sorting:
- GPT-jax based on the official huggingface library☆13Updated 4 years ago
- Training and evaluation code for the paper "Headless Language Models: Learning without Predicting with Contrastive Weight Tying" (https:/…☆27Updated last year
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated 2 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆50Updated 3 years ago
- Implementation of N-Grammer, augmenting Transformers with latent n-grams, in Pytorch☆76Updated 2 years ago
- Official code for "Distributed Deep Learning in Open Collaborations" (NeurIPS 2021)☆117Updated 3 years ago
- Tutorial to pretrain & fine-tune a 🤗 Flax T5 model on a TPUv3-8 with GCP☆58Updated 3 years ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pile☆116Updated 2 years ago
- [COLM 2024] Early Weight Averaging meets High Learning Rates for LLM Pre-training☆17Updated 11 months ago
- ☆13Updated 6 years ago
- A minimal PyTorch Lightning OpenAI GPT w DeepSpeed Training!☆113Updated 2 years ago
- c++ mosestokenizer☆18Updated last year
- A boilerplate to use multiprocessing for your gRPC server in your Python project☆26Updated 4 years ago
- Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch☆59Updated 4 years ago
- sigma-MoE layer☆20Updated last year
- High performance pytorch modules☆18Updated 2 years ago
- ☆13Updated 9 months ago
- Truly flash T5 realization!☆70Updated last year
- My explorations into editing the knowledge and memories of an attention network☆35Updated 2 years ago
- Large Scale Distributed Model Training strategy with Colossal AI and Lightning AI☆56Updated 2 years ago
- Hidden Engrams: Long Term Memory for Transformer Model Inference☆35Updated 4 years ago
- This repository contains example code to build models on TPUs☆30Updated 2 years ago
- Repo for training MLMs, CLMs, or T5-type models on the OLM pretraining data, but it should work with any hugging face text dataset.☆95Updated 2 years ago
- Interactive Neural Machine Translation-lite (INMT-lite) is a framework to train and develop lite versions (.tflite) of models for neural …☆47Updated last year
- Experiments for XLM-V Transformers Integeration☆13Updated 2 years ago
- Online (real-time) decoder to be used with DeepSpeech2 model☆25Updated 5 years ago
- Load What You Need: Smaller Multilingual Transformers for Pytorch and TensorFlow 2.0.☆105Updated 3 years ago
- Experiments with generating opensource language model assistants☆97Updated 2 years ago
- Running massive simulations using RNNs on CPUs for building bots and all kinds of things.☆14Updated 4 years ago
- Helper scripts and notes that were used while porting various nlp models☆47Updated 3 years ago