kjslag / spacebyteLinks
A byte-level decoder architecture that matches the performance of tokenized Transformers.
☆66Updated last year
Alternatives and similar repositories for spacebyte
Users that are interested in spacebyte are comparing it to the libraries listed below
Sorting:
- Griffin MQA + Hawk Linear RNN Hybrid☆89Updated last year
- Pytorch implementation of the PEER block from the paper, Mixture of A Million Experts, by Xu Owen He at Deepmind☆131Updated last week
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆101Updated last year
- RWKV-7: Surpassing GPT☆98Updated 11 months ago
- PyTorch implementation of models from the Zamba2 series.☆185Updated 9 months ago
- ☆87Updated last year
- Tiled Flash Linear Attention library for fast and efficient mLSTM Kernels.☆73Updated 3 weeks ago
- EvaByte: Efficient Byte-level Language Models at Scale☆110Updated 6 months ago
- Collection of autoregressive model implementation☆86Updated 6 months ago
- ☆39Updated last year
- Token Omission Via Attention☆127Updated last year
- ☆81Updated last year
- ☆48Updated last year
- ☆52Updated last year
- Experimental playground for benchmarking language model (LM) architectures, layers, and tricks on smaller datasets. Designed for flexible…☆86Updated this week
- An unofficial pytorch implementation of 'Efficient Infinite Context Transformers with Infini-attention'☆54Updated last year
- GoldFinch and other hybrid transformer components☆45Updated last year
- Q-GaLore: Quantized GaLore with INT4 Projection and Layer-Adaptive Low-Rank Gradients.☆202Updated last year
- RWKV, in easy to read code☆72Updated 7 months ago
- Block Transformer: Global-to-Local Language Modeling for Fast Inference (NeurIPS 2024)☆162Updated 7 months ago
- Mixture of A Million Experts☆49Updated last year
- Fast, Modern, and Low Precision PyTorch Optimizers☆116Updated 2 months ago
- ☆57Updated last month
- A repository for research on medium sized language models.☆78Updated last year
- This is the official repository for Inheritune.☆115Updated 9 months ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆38Updated 5 months ago
- Implementation of the Mamba SSM with hf_integration.☆56Updated last year
- ☆50Updated last year
- Code for Zero-Shot Tokenizer Transfer☆140Updated 9 months ago
- ☆65Updated 7 months ago