RuslanKhalitov / ChordMixer
The official implementation of the ChordMixer architecture.
☆61Updated last year
Alternatives and similar repositories for ChordMixer:
Users that are interested in ChordMixer are comparing it to the libraries listed below
- Official code for Long Expressive Memory (ICLR 2022, Spotlight)☆69Updated 2 years ago
- FusionBrain Challenge 2.0: creating multimodal multitask model☆16Updated 2 years ago
- Compression schema for gradients of activations in backward pass☆44Updated last year
- ☆71Updated 5 months ago
- ☆17Updated 2 months ago
- ☆21Updated last year
- Deep Learning Audio Course – AI Masters☆27Updated 9 months ago
- Code for MSID, a Multi-Scale Intrinsic Distance for comparing generative models, studying neural networks, and more!☆51Updated 5 years ago
- Framework for processing and filtering datasets☆27Updated 6 months ago
- Why Do We Need Weight Decay in Modern Deep Learning? [NeurIPS 2024]☆60Updated 4 months ago
- ☆20Updated 6 months ago
- Deep Generative Models course, 2021☆21Updated 3 years ago
- A repo based on XiLin Li's PSGD repo that extends some of the experiments.☆14Updated 4 months ago
- This is the official repo for Gradient Agreement Filtering (GAF).☆22Updated 3 weeks ago
- DiCE: The Infinitely Differentiable Monte-Carlo Estimator☆31Updated last year
- [ICML 2024] SIRFShampoo: Structured inverse- and root-free Shampoo in PyTorch (https://arxiv.org/abs/2402.03496)☆14Updated 3 months ago
- Inspired by "Neural Networks Fail to Learn Periodic Functions and How to Fix It"☆62Updated 9 months ago
- GULAG: GUessing LAnGuages with neural networks☆13Updated 2 years ago
- Implementation of "Gradients without backpropagation" paper (https://arxiv.org/abs/2202.08587) using functorch☆108Updated last year
- Simple audio AE☆12Updated 3 months ago
- Implementation of GateLoop Transformer in Pytorch and Jax☆87Updated 8 months ago
- AdaCat☆49Updated 2 years ago
- ☆36Updated last year
- Implementation of a Light Recurrent Unit in Pytorch☆48Updated 4 months ago
- RuCLIP tiny (Russian Contrastive Language–Image Pretraining) is a neural network trained to work with different pairs (images, texts).☆32Updated 2 years ago
- Exploring an idea where one forgets about efficiency and carries out attention across each edge of the nodes (tokens)☆44Updated this week
- Codes accompanying the paper "LaProp: a Better Way to Combine Momentum with Adaptive Gradient"☆27Updated 4 years ago
- Code for the paper "PALBERT: Teaching ALBERT to Ponder", NeurIPS 2022 Spotlight☆37Updated last year
- Skoltech NLA 2024 course.☆25Updated 2 months ago
- ☆13Updated 3 years ago