quinte22 / bumblebeeLinks
bumble bee transformer
☆14Updated 4 years ago
Alternatives and similar repositories for bumblebee
Users that are interested in bumblebee are comparing it to the libraries listed below
Sorting:
- Implementation of N-Grammer, augmenting Transformers with latent n-grams, in Pytorch☆76Updated 3 years ago
- Implementation of Multistream Transformers in Pytorch☆54Updated 4 years ago
- Anonymous ICLR Submission☆14Updated 6 years ago
- A deep learning library based on Pytorch focussed on low resource language research and robustness☆70Updated 4 years ago
- Enable RNNLM lattice rescoring with Pytorch [kaldi]☆12Updated 5 years ago
- Prabhupadavani: A Code-mixed Speech Translation Data for 25 languages☆13Updated 3 years ago
- Implementation of COCO-LM, Correcting and Contrasting Text Sequences for Language Model Pretraining, in Pytorch☆46Updated 4 years ago
- A python library for highly configurable transformers - easing model architecture search and experimentation.☆49Updated 4 years ago
- High performance pytorch modules☆18Updated 2 years ago
- 🤗 Transformers: State-of-the-art Machine Learning for Pytorch, TensorFlow, and JAX.☆81Updated 3 years ago
- Contrastive Language-Audio Pretraining☆15Updated 4 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- Code base for WaveTransformer: A novel architecture for automated audio captioning☆44Updated 4 years ago
- Repository with illustrations for cft-contest-2018☆12Updated 7 years ago
- ☆27Updated 6 years ago
- Source code for "A Lightweight Recurrent Network for Sequence Modeling"☆26Updated 3 years ago
- Local Attention - Flax module for Jax☆22Updated 4 years ago
- This is a intuitive explanation of Representation Learning with Contrastive Predictive Coding using code provided by jefflai108 that use…☆10Updated 4 years ago
- Code for "MIM: Mutual Information Machine" paper.☆15Updated 3 years ago
- PyTorch implementation of the paper "NanoFlow: Scalable Normalizing Flows with Sublinear Parameter Complexity." (NeurIPS 2020)☆66Updated 5 years ago
- Unofficial PyTorch implementation of Fastformer based on paper "Fastformer: Additive Attention Can Be All You Need"."☆132Updated 4 years ago
- some tutorials for blog: simonjisu.github.io☆23Updated 4 years ago
- A punctuation transcription model to automatically add punctuation marks in an unpunctuated sentence or sentences.☆15Updated 5 years ago
- A Benchmark Dataset for Understanding Disfluencies in Question Answering☆64Updated 4 years ago
- PyTorch reimplementation of the paper "HyperMixer: An MLP-based Green AI Alternative to Transformers" [arXiv 2022].☆18Updated 3 years ago
- Online (real-time) decoder to be used with DeepSpeech2 model☆25Updated 5 years ago
- Large Scale BERT Distillation☆33Updated 2 years ago
- Factorization of the neural parameter space for zero-shot multi-lingual and multi-task transfer☆39Updated 5 years ago
- ☆76Updated 4 years ago
- Code repo for "Transformer on a Diet" paper☆31Updated 5 years ago