Whiax / BERT-Transformer-Pytorch
Basic implementation of BERT and Transformer in Pytorch in one short python file (also includes "predict next word" GPT task)
☆42Updated last year
Alternatives and similar repositories for BERT-Transformer-Pytorch:
Users that are interested in BERT-Transformer-Pytorch are comparing it to the libraries listed below
- An implementation of masked language modeling for Pytorch, made as concise and simple as possible☆180Updated last year
- Some notebooks for NLP☆200Updated last year
- Code for "Finetuning Pretrained Transformers into Variational Autoencoders"☆39Updated 2 years ago
- Implementation of the GBST block from the Charformer paper, in Pytorch☆116Updated 3 years ago
- A library for making Transformer Variational Autoencoders. (Extends the Huggingface/transformers library.)☆138Updated 3 years ago
- ☆48Updated 2 years ago
- Checkout the new version at the link!☆22Updated 4 years ago
- A simple and working implementation of Electra, the fastest way to pretrain language models from scratch, in Pytorch☆226Updated last year
- ☆46Updated 4 years ago
- ☆64Updated 4 years ago
- Pretrain and finetune ELECTRA with fastai and huggingface. (Results of the paper replicated !)☆328Updated last year
- The code for the video tutorial series on building a Transformer from scratch: https://www.youtube.com/watch?v=XR4VDnJzB8o☆18Updated 2 years ago
- Language Modeling Example with Transformers and PyTorch Lighting☆65Updated 4 years ago
- ☆206Updated last year
- this is where we share notebooks/projects used in your youtube channel☆148Updated 3 years ago
- Implementation of Feedback Transformer in Pytorch☆105Updated 4 years ago
- A PyTorch implementation of the Transformer model from "Attention Is All You Need".☆59Updated 5 years ago
- Trains Transformer model variants. Data isn't shuffled between batches.☆142Updated 2 years ago
- Accelerated NLP pipelines for fast inference on CPU. Built with Transformers and ONNX runtime.☆126Updated 4 years ago
- NLP Examples using the 🤗 libraries☆41Updated 4 years ago
- Sequence to Sequence Models in PyTorch☆44Updated 8 months ago
- Well documented, unit tested, type checked and formatted implementation of a vanilla transformer - for educational purposes.☆242Updated last year
- Implementation of H-Transformer-1D, Hierarchical Attention for Sequence Learning☆160Updated last year
- Code for the Shortformer model, from the ACL 2021 paper by Ofir Press, Noah A. Smith and Mike Lewis.☆147Updated 3 years ago
- Central repository for all lectures on deep learning at UPC ETSETB TelecomBCN.☆53Updated 2 years ago
- Tensorflow implementation of a linear attention architecture☆44Updated 4 years ago
- http://nlp.seas.harvard.edu/2018/04/03/attention.html☆62Updated 3 years ago
- ☆34Updated 4 years ago
- A 🤗-style implementation of BERT using lambda layers instead of self-attention☆69Updated 4 years ago
- Tutorial for first time BERT users,☆103Updated 2 years ago