facebookresearch / QincoLinks
Residual Quantization with Implicit Neural Codebooks
☆105Updated last month
Alternatives and similar repositories for Qinco
Users that are interested in Qinco are comparing it to the libraries listed below
Sorting:
- [ICLR 2025 & COLM 2025] Official PyTorch implementation of the Forgetting Transformer and Adaptive Computation Pruning☆134Updated 3 weeks ago
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆131Updated 3 weeks ago
- Implementation of the proposed DeepCrossAttention by Heddes et al at Google research, in Pytorch☆94Updated 9 months ago
- Triton implement of bi-directional (non-causal) linear attention☆56Updated 9 months ago
- Implementation of the dynamic chunking mechanism in H-net by Hwang et al. of Carnegie Mellon☆65Updated 3 months ago
- Implementation of the proposed Adam-atan2 from Google Deepmind in Pytorch☆134Updated last month
- Implementation of a Light Recurrent Unit in Pytorch☆49Updated last year
- Implementation of a multimodal diffusion transformer in Pytorch☆106Updated last year
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆162Updated 7 months ago
- Implementation of Qformer from BLIP2 in Zeta Lego blocks.☆46Updated last year
- A practical implementation of GradNorm, Gradient Normalization for Adaptive Loss Balancing, in Pytorch☆118Updated 3 months ago
- HGRN2: Gated Linear RNNs with State Expansion☆55Updated last year
- Explorations into the recently proposed Taylor Series Linear Attention☆100Updated last year
- Attempt to make multiple residual streams from Bytedance's Hyper-Connections paper accessible to the public☆93Updated 5 months ago
- Official Code for Paper: Beyond Matryoshka: Revisiting Sparse Coding for Adaptive Representation☆128Updated 4 months ago
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best…☆54Updated 8 months ago
- [ICML 2024] When Linear Attention Meets Autoregressive Decoding: Towards More Effective and Efficient Linearized Large Language Models☆36Updated last year
- ResiDual: Transformer with Dual Residual Connections, https://arxiv.org/abs/2304.14802☆96Updated 2 years ago
- Code repository for the public reproduction of the language modelling experiments on "MatFormer: Nested Transformer for Elastic Inference…☆30Updated 2 years ago
- Experimental playground for benchmarking language model (LM) architectures, layers, and tricks on smaller datasets. Designed for flexible…☆87Updated 2 weeks ago
- User-friendly implementation of the Mixture-of-Sparse-Attention (MoSA). MoSA selects distinct tokens for each head with expert choice rou…☆28Updated 6 months ago
- Distributed Optimization Infra for learning CLIP models☆27Updated last year
- ☆96Updated 9 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆68Updated last year
- ☆154Updated 7 months ago
- Pytorch Implementation of the sparse attention from the paper: "Generating Long Sequences with Sparse Transformers"☆93Updated last month
- Official implementation of the paper: "ZClip: Adaptive Spike Mitigation for LLM Pre-Training".☆139Updated last week
- Flash Attention Triton kernel with support for second-order derivatives☆115Updated last month
- Beyond Straight-Through☆105Updated 2 years ago
- Implementation of Agent Attention in Pytorch☆92Updated last year