lucidrains / holodeck-pytorchView external linksLinks
Implementation of a holodeck, written in Pytorch
☆18Nov 1, 2023Updated 2 years ago
Alternatives and similar repositories for holodeck-pytorch
Users that are interested in holodeck-pytorch are comparing it to the libraries listed below
Sorting:
- Implementation of a simple BPE tokenizer, but in Nim☆22Jul 2, 2023Updated 2 years ago
- A simple Transformer where the softmax has been replaced with normalization☆20Sep 11, 2020Updated 5 years ago
- ☆11Oct 11, 2023Updated 2 years ago
- Implementation of an Attention layer where each head can attend to more than just one token, using coordinate descent to pick topk☆47Jul 16, 2023Updated 2 years ago
- Implementation of Insertion-deletion Denoising Diffusion Probabilistic Models☆30May 31, 2022Updated 3 years ago
- ☆13Jun 3, 2024Updated last year
- Thispersondoesnotexist went down, so this time, while building it back up, I am going to open source all of it.☆91Aug 26, 2023Updated 2 years ago
- Implementation of Discrete Key / Value Bottleneck, in Pytorch☆88Jul 9, 2023Updated 2 years ago
- Implementation of GateLoop Transformer in Pytorch and Jax☆92Jun 18, 2024Updated last year
- Implementation and explorations into Blackbox Gradient Sensing (BGS), an evolutionary strategies approach proposed in a Google Deepmind p…☆20Jul 20, 2025Updated 6 months ago
- My own attempt at a long context genomics model, leveraging recent advances in long context attention modeling (Flash Attention + other h…☆54Jul 2, 2023Updated 2 years ago
- An attempt to merge ESBN with Transformers, to endow Transformers with the ability to emergently bind symbols☆16Aug 3, 2021Updated 4 years ago
- Visualising Losses in Deep Neural Networks☆16Jul 17, 2024Updated last year
- Implementation of Memory-Compressed Attention, from the paper "Generating Wikipedia By Summarizing Long Sequences"☆70Apr 10, 2023Updated 2 years ago
- Latent Diffusion Language Models☆70Sep 20, 2023Updated 2 years ago
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆46May 23, 2023Updated 2 years ago
- Implementation of RQ Transformer, proposed in the paper "Autoregressive Image Generation using Residual Quantization"☆124Apr 19, 2022Updated 3 years ago
- Local Attention - Flax module for Jax☆22May 26, 2021Updated 4 years ago
- Implementation of Metaformer, but in an autoregressive manner☆26Jun 21, 2022Updated 3 years ago
- Standalone Product Key Memory module in Pytorch - for augmenting Transformer models☆87Nov 1, 2025Updated 3 months ago
- ImageNet-12k subset of ImageNet-21k (fall11)☆21Jun 13, 2023Updated 2 years ago
- ☆27Jul 28, 2025Updated 6 months ago
- Implementation of a Light Recurrent Unit in Pytorch☆49Oct 6, 2024Updated last year
- My attempts at applying Soundstream design on learned tokenization of text and then applying hierarchical attention to text generation☆90Oct 11, 2024Updated last year
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)☆190Jun 24, 2022Updated 3 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Apr 6, 2022Updated 3 years ago
- ☆22Jul 7, 2023Updated 2 years ago
- Utilities for PyTorch distributed☆25Feb 27, 2025Updated 11 months ago
- Train vision models using JAX and 🤗 transformers☆100Dec 14, 2025Updated 2 months ago
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆55Mar 25, 2025Updated 10 months ago
- ☆24Sep 25, 2024Updated last year
- How to use tensorboard in fastai☆21Jul 10, 2019Updated 6 years ago
- Implementation of Nvidia's NeuralPlexer, for end-to-end differentiable design of functional small-molecules and ligand-binding proteins, …☆52Nov 20, 2023Updated 2 years ago
- Implementation of Hourglass Transformer, in Pytorch, from Google and OpenAI☆98Dec 31, 2021Updated 4 years ago
- Reference implementation of "Softmax Attention with Constant Cost per Token" (Heinsen, 2024)☆24Jun 6, 2024Updated last year
- Yet another random morning idea to be quickly tried and architecture shared if it works; to allow the transformer to pause for any amount…☆53Oct 22, 2023Updated 2 years ago
- Implementation of Multistream Transformers in Pytorch☆54Jul 31, 2021Updated 4 years ago
- Exploration into the proposed "Self Reasoning Tokens" by Felipe Bonetto☆57May 17, 2024Updated last year
- Stick-breaking attention☆62Jul 1, 2025Updated 7 months ago