DACUS1995 / pytorch-mmap-datasetLinks
A custom pytorch Dataset extension that provides a faster iteration and better RAM usage
☆43Updated last year
Alternatives and similar repositories for pytorch-mmap-dataset
Users that are interested in pytorch-mmap-dataset are comparing it to the libraries listed below
Sorting:
- A Tight-fisted Optimizer☆48Updated 2 years ago
- PyTorch, PyTorch Lightning framework for trying knowledge distillation in image classification problems☆32Updated 11 months ago
- The accompanying code for "Memory-efficient Transformers via Top-k Attention" (Ankit Gupta, Guy Dar, Shaya Goodman, David Ciprut, Jonatha…☆69Updated 3 years ago
- Pytorch cyclic cosine decay learning rate scheduler☆48Updated 4 years ago
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆39Updated 3 years ago
- Lion and Adam optimization comparison☆61Updated 2 years ago
- PyTorch implementation of Soft MoE by Google Brain in "From Sparse to Soft Mixtures of Experts" (https://arxiv.org/pdf/2308.00951.pdf)☆73Updated last year
- several types of attention modules written in PyTorch for learning purposes☆52Updated 8 months ago
- Implementation of IceFormer: Accelerated Inference with Long-Sequence Transformers on CPUs (ICLR 2024).☆25Updated last year
- A repository for DenseSSMs☆87Updated last year
- [NeurIPS 2022 Spotlight] This is the official PyTorch implementation of "EcoFormer: Energy-Saving Attention with Linear Complexity"☆72Updated 2 years ago
- Implementation of a Transformer using ReLA (Rectified Linear Attention) from https://arxiv.org/abs/2104.07012☆49Updated 3 years ago
- Skyformer: Remodel Self-Attention with Gaussian Kernel and Nystr\"om Method (NeurIPS 2021)☆61Updated 3 years ago
- [ICLR 2022] Official implementation of cosformer-attention in cosFormer: Rethinking Softmax in Attention☆193Updated 2 years ago
- Warmup learning rate wrapper for Pytorch Scheduler☆41Updated 5 years ago
- 32 times longer context window than vanilla Transformers and up to 4 times longer than memory efficient Transformers.☆48Updated 2 years ago
- ☆31Updated last year
- Implementation of a Light Recurrent Unit in Pytorch☆48Updated 8 months ago
- Implementation of Infini-Transformer in Pytorch☆111Updated 5 months ago
- CUDA implementation of autoregressive linear attention, with all the latest research findings☆44Updated 2 years ago
- [EMNLP 2022] Official implementation of Transnormer in our EMNLP 2022 paper - The Devil in Linear Transformer☆60Updated last year
- ☆8Updated last year
- [ICLR 2022] "As-ViT: Auto-scaling Vision Transformers without Training" by Wuyang Chen, Wei Huang, Xianzhi Du, Xiaodan Song, Zhangyang Wa…☆76Updated 3 years ago
- PyTorch implementation of Train Short, Test Long: Attention with Linear Biases Enables Input Length Extrapolation☆28Updated 3 years ago
- # Unified Normalization (ACM MM'22) By Qiming Yang, Kai Zhang, Chaoxiang Lan, Zhi Yang, Zheyang Li, Wenming Tan, Jun Xiao, and Shiliang P…☆34Updated 2 years ago
- A simple program to calculate and visualize the FLOPs and Parameters of Pytorch models, with handy CLI and easy-to-use Python API.☆129Updated 7 months ago
- A torch-based implementation of K-Means and K-Means++☆17Updated 4 years ago
- A PyTorch Lightning extension that accelerates and enhances foundation model experimentation with flexible fine-tuning schedules.☆63Updated 2 months ago
- ☆16Updated 2 years ago
- ☆16Updated 2 years ago