pbloem / former
Simple transformer implementation from scratch in pytorch.
☆1,068Updated 8 months ago
Alternatives and similar repositories for former:
Users that are interested in former are comparing it to the libraries listed below
- Fast, general, and tested differentiable structured prediction in PyTorch☆1,109Updated 2 years ago
- My implementation of the original transformer model (Vaswani et al.). I've additionally included the playground.py file for visualizing o…☆1,012Updated 4 years ago
- ☆779Updated 10 months ago
- Hopfield Networks is All You Need☆1,765Updated last year
- Reformer, the efficient Transformer, in Pytorch☆2,144Updated last year
- Pytorch library for fast transformer implementations☆1,670Updated last year
- Transformer training code for sequential tasks☆609Updated 3 years ago
- PyTorch implementation of some attentions for Deep Learning Researchers.☆520Updated 2 years ago
- Examples of using sparse attention, as in "Generating Long Sequences with Sparse Transformers"☆1,544Updated 4 years ago
- Toolbox of models, callbacks, and datasets for AI/ML researchers.☆1,709Updated 3 weeks ago
- VQVAEs, GumbelSoftmaxes and friends☆549Updated 3 years ago
- A Unified Library for Parameter-Efficient and Modular Transfer Learning☆2,640Updated this week
- Implementation of Transformer Model in Tensorflow☆465Updated last year
- Long Range Arena for Benchmarking Efficient Transformers☆740Updated last year
- A repository containing tutorials for practical NLP using PyTorch☆532Updated 5 years ago
- Fast Block Sparse Matrices for Pytorch☆546Updated 4 years ago
- An implementation of Performer, a linear attention-based transformer, in Pytorch☆1,114Updated 2 years ago
- Flexible components pairing 🤗 Transformers with Pytorch Lightning☆612Updated 2 years ago
- Course notes☆653Updated 9 months ago
- An unofficial styleguide and best practices summary for PyTorch☆1,952Updated 3 years ago
- arxiv-sanity lite: tag arxiv papers of interest get recommendations of similar papers in a nice UI using SVMs over tfidf feature vectors …☆1,212Updated last year
- Pytorch Lightning code guideline for conferences☆1,245Updated last year
- Helps you write algorithms in PyTorch that adapt to the available (CUDA) memory☆431Updated 5 months ago
- Transformers for Longer Sequences☆586Updated 2 years ago
- PyTorch tutorials and best practices.☆1,675Updated 2 years ago
- PyTorch extensions for high performance and large scale training.☆3,238Updated 2 weeks ago
- Pre-trained subword embeddings in 275 languages, based on Byte-Pair Encoding (BPE)☆1,193Updated 3 months ago
- A collection of resources to study Transformers in depth.☆522Updated last year
- Model interpretability and understanding for PyTorch☆5,052Updated last week
- Transformer based on a variant of attention that is linear complexity in respect to sequence length☆731Updated 8 months ago