UW-Madison-Lee-Lab / Expressive_Power_of_LoRALinks
Code for "The Expressive Power of Low-Rank Adaptation".
☆20Updated last year
Alternatives and similar repositories for Expressive_Power_of_LoRA
Users that are interested in Expressive_Power_of_LoRA are comparing it to the libraries listed below
Sorting:
- ☆34Updated 2 years ago
- ☆30Updated 2 years ago
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆27Updated last year
- Code Release for "Broken Neural Scaling Laws" (BNSL) paper☆59Updated 2 years ago
- ☆103Updated last year
- ☆20Updated last year
- Code for the paper "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns"☆18Updated last year
- ☆12Updated last year
- Self-Supervised Alignment with Mutual Information☆21Updated last year
- Experiments and code to generate the GINC small-scale in-context learning dataset from "An Explanation for In-context Learning as Implici…☆106Updated last year
- ☆19Updated 11 months ago
- ☆45Updated 2 years ago
- Companion repository to "Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models"☆14Updated 2 years ago
- ☆20Updated last year
- Code for the paper: https://arxiv.org/pdf/2309.06979.pdf☆21Updated last year
- Blog post☆17Updated last year
- Code for reproducing our paper "Low Rank Adapting Models for Sparse Autoencoder Features"☆17Updated 6 months ago
- [ACL 2023]: Training Trajectories of Language Models Across Scales https://arxiv.org/pdf/2212.09803.pdf☆25Updated last year
- ZeroC is a neuro-symbolic method that trained with elementary visual concepts and relations, can zero-shot recognize and acquire more com…☆32Updated 2 years ago
- ☆16Updated 2 years ago
- A Kernel-Based View of Language Model Fine-Tuning https://arxiv.org/abs/2210.05643☆78Updated 2 years ago
- Curse-of-memory phenomenon of RNNs in sequence modelling☆18Updated 5 months ago
- ☆32Updated 7 months ago
- ☆46Updated 2 years ago
- Exploration of automated dataset selection approaches at large scales.☆48Updated 7 months ago
- ☆16Updated last year
- Efficient Scaling laws and collaborative pretraining.☆18Updated last month
- Codebase for Context-aware Meta-learned Loss Scaling (CaMeLS). https://arxiv.org/abs/2305.15076.☆25Updated last year
- Bayesian Low-Rank Adaptation for Large Language Models☆36Updated last year
- ☆31Updated last year