sc782 / SBM-Transformer
☆13Updated last year
Related projects ⓘ
Alternatives and complementary repositories for SBM-Transformer
- Official Code for Local Search GFlowNets (ICLR 2024 Spotlight)☆15Updated 2 months ago
- Official PyTorch implementation of "Energy-Based Contrastive Learning of Visual Representations", NeurIPS 2022 Oral Paper☆9Updated 2 years ago
- ☆21Updated 2 years ago
- Code for GFlowNet-EM, a novel algorithm for fitting latent variable models with compositional latents and an intractable true posterior.☆38Updated 9 months ago
- [NeurIPS'23 Spotlight] Learning Probabilistic Symmetrization for Architecture Agnostic Equivariance (LPS), in PyTorch☆26Updated 6 months ago
- Pytorch implementation of neural processes and variants☆26Updated 3 months ago
- ☆25Updated 2 weeks ago
- ☆75Updated last year
- ☆14Updated last year
- Position Prediction as an Effective Pretraining Strategy☆8Updated last year
- [NeurIPS'21] Higher-order Transformers for sets, graphs, and hypergraphs, in PyTorch☆60Updated last year
- Bayesian Attention Modules☆35Updated 3 years ago
- Repository for "Generative Flow Networks as Entropy-Regularized RL" (AISTATS-2024, Oral)☆24Updated 6 months ago
- Code for our TMLR paper "Distributional GFlowNets with Quantile Flows".☆10Updated 8 months ago
- ☆14Updated last year
- [ICLR 2024 Oral] Beyond Weisfeiler-Lehman: A Quantitative Framework for GNN Expressiveness.☆15Updated 9 months ago
- Online Adaptation of Language Models with a Memory of Amortized Contexts (NeurIPS 2024)☆53Updated 3 months ago
- Curse-of-memory phenomenon of RNNs in sequence modelling☆19Updated this week
- Code for "Can We Scale Transformers to Predict Parameters of Diverse ImageNet Models?" [ICML 2023]☆31Updated 2 months ago
- Official PyTorch implementation of NPwSA: "Neural Processes with Stochastic Attention: Paying more attention to the context dataset (ICLR…☆10Updated 2 years ago
- [NeurIPS 2022] Your Transformer May Not be as Powerful as You Expect (official implementation)☆33Updated last year
- Code for the paper "What Makes Better Augmentation Strategies? Augment Difficult but Not too Different" (ICLR 22)☆13Updated last year
- ☆13Updated 6 months ago
- Code to reproduce the results for Compositional Attention☆60Updated last year
- ☆31Updated 10 months ago
- ☆50Updated last year
- MambaFormer in-context learning experiments and implementation for https://arxiv.org/abs/2402.04248☆34Updated 4 months ago
- ☆10Updated last year
- ☆36Updated 3 years ago
- ☆28Updated this week