sradc / pretraining-BERT
Pre-train BERT from scratch, with HuggingFace. Accompanies the blog post: sidsite.com/posts/bert-from-scratch
☆37Updated last year
Related projects: ⓘ
- A lightweight PyTorch implementation of the Transformer-XL architecture proposed by Dai et al. (2019)