tt-embedding / tt-embeddingsLinks
☆28Updated 6 years ago
Alternatives and similar repositories for tt-embeddings
Users that are interested in tt-embeddings are comparing it to the libraries listed below
Sorting:
- Pytorch library for factorized L0-based pruning.☆45Updated 2 years ago
- ☆59Updated 5 years ago
- Compression of NMT transformer model with tensor methods☆48Updated 6 years ago
- u-MPS implementation and experimentation code used in the paper Tensor Networks for Probabilistic Sequence Modeling (https://arxiv.org/ab…☆19Updated 5 years ago
- [ICLR 2022] Code for paper "Exploring Extreme Parameter Compression for Pre-trained Language Models"(https://arxiv.org/abs/2205.10036)☆22Updated 2 years ago
- ☆64Updated 5 years ago
- Differentiable Product Quantization for End-to-End Embedding Compression.☆64Updated 3 years ago
- PyTorch implementation for PaLM: A Hybrid Parser and Language Model.☆10Updated 5 years ago
- ☆33Updated 4 years ago
- Codes for "Understanding and Improving Transformer From a Multi-Particle Dynamic System Point of View"☆147Updated 6 years ago
- This package implements THOR: Transformer with Stochastic Experts.☆65Updated 4 years ago
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643☆78Updated 2 years ago
- Block Sparse movement pruning☆81Updated 5 years ago
- Code for the paper "A Theoretical Analysis of the Repetition Problem in Text Generation" in AAAI 2021.☆57Updated 3 years ago
- [NeurIPS 2020] "The Lottery Ticket Hypothesis for Pre-trained BERT Networks", Tianlong Chen, Jonathan Frankle, Shiyu Chang, Sijia Liu, Ya…☆141Updated 4 years ago
- Implementation of NeurIPS 20 paper: Latent Template Induction with Gumbel-CRFs☆56Updated 5 years ago
- Source code of paper "BP-Transformer: Modelling Long-Range Context via Binary Partitioning"☆128Updated 4 years ago
- An implementation of various tensor-based decomposition for NN & RNN parameters☆18Updated 7 years ago
- Code for the paper "BERT Loses Patience: Fast and Robust Inference with Early Exit".☆66Updated 4 years ago
- Source code for "Efficient Training of BERT by Progressively Stacking"☆113Updated 6 years ago
- Rationales for Sequential Predictions☆40Updated 3 years ago
- [EMNLP'19] Summary for Transformer Understanding☆53Updated 6 years ago
- Code for the paper "Are Sixteen Heads Really Better than One?"☆175Updated 5 years ago
- ☆44Updated 7 years ago
- ☆13Updated 4 years ago
- Method to improve inference time for BERT. This is an implementation of the paper titled "PoWER-BERT: Accelerating BERT Inference via Pro…☆62Updated 3 months ago
- Blog post☆17Updated last year
- [ACL‘20] Highway Transformer: A Gated Transformer.☆33Updated 4 years ago
- A simple module consistently outperforms self-attention and Transformer model on main NMT datasets with SoTA performance.☆86Updated 2 years ago
- ☆14Updated 6 years ago