jungokasai / T2RLinks
☆14Updated 2 years ago
Alternatives and similar repositories for T2R
Users that are interested in T2R are comparing it to the libraries listed below
Sorting:
- triton ver of gqa flash attn, based on the tutorial☆12Updated last year
- Official Implementation of ACL2023: Don't Parse, Choose Spans! Continuous and Discontinuous Constituency Parsing via Autoregressive Span …☆13Updated last year
- ☆20Updated last year
- A probabilitic model for contextual word representation. Accepted to ACL2023 Findings.☆24Updated last year
- ☆31Updated 2 years ago
- This repository is the official implementation of our EMNLP 2022 paper ELMER: A Non-Autoregressive Pre-trained Language Model for Efficie…☆26Updated 2 years ago
- A Structured Span Selector (NAACL 2022). A structured span selector with a WCFG for span selection tasks (coreference resolution, semanti…☆21Updated 3 years ago
- A method for evaluating the high-level coherence of machine-generated texts. Identifies high-level coherence issues in transformer-based …☆11Updated 2 years ago
- [ICML 2023] "Data Efficient Neural Scaling Law via Model Reusing" by Peihao Wang, Rameswar Panda, Zhangyang Wang☆14Updated last year
- Staged Training for Transformer Language Models☆32Updated 3 years ago
- [EMNLP 2022] Language Model Pre-Training with Sparse Latent Typing☆14Updated 2 years ago
- ☆13Updated 2 years ago
- ☆55Updated last year
- ☆19Updated 2 years ago
- Contextual Position Encoding but with some custom CUDA Kernels https://arxiv.org/abs/2405.18719☆22Updated last year
- Fine-Tuning Pre-trained Transformers into Decaying Fast Weights☆19Updated 2 years ago
- PyTorch implementation for PaLM: A Hybrid Parser and Language Model.☆10Updated 5 years ago
- Use the tokenizer in parallel to achieve superior acceleration☆17Updated last year
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆38Updated last year
- Implementation of the model: "Reka Core, Flash, and Edge: A Series of Powerful Multimodal Language Models" in PyTorch☆30Updated 2 weeks ago
- ☆15Updated 2 years ago
- Transformers at any scale☆41Updated last year
- [NeurIPS 2023] Repetition In Repetition Out: Towards Understanding Neural Text Degeneration from the Data Perspective☆33Updated last year
- [ICLR 2022] Code for paper "Exploring Extreme Parameter Compression for Pre-trained Language Models"(https://arxiv.org/abs/2205.10036)☆22Updated 2 years ago
- ☆22Updated last year
- The source code of "Merging Experts into One: Improving Computational Efficiency of Mixture of Experts (EMNLP 2023)":☆38Updated last year
- This package implements THOR: Transformer with Stochastic Experts.☆65Updated 3 years ago
- The official implementation for Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-Free☆46Updated 2 months ago
- Code for the preprint "Cache Me If You Can: How Many KVs Do You Need for Effective Long-Context LMs?"☆41Updated last week
- Xmixers: A collection of SOTA efficient token/channel mixers☆11Updated last month