quanta-fine-tuning / quantaLinks
(NeurIPS 2024) QuanTA: Efficient High-Rank Fine-Tuning of LLMs with Quantum-Informed Tensor Adaptation
☆30Updated 8 months ago
Alternatives and similar repositories for quanta
Users that are interested in quanta are comparing it to the libraries listed below
Sorting:
- Code for the paper: Why Transformers Need Adam: A Hessian Perspective☆60Updated 4 months ago
- A thoroughly investigated survey for tensorial neural networks.☆137Updated 6 months ago
- Omnigrok: Grokking Beyond Algorithmic Data☆61Updated 2 years ago
- Collect optimizer related papers, data, repositories☆94Updated 8 months ago
- Unofficial Implementation of Selective Attention Transformer☆17Updated 9 months ago
- [NeurIPS 2023 Spotlight] Temperature Balancing, Layer-wise Weight Analysis, and Neural Network Training☆35Updated 4 months ago
- This repository contains the official code for Energy Transformer---an efficient Energy-based Transformer variant for graph classificatio…☆25Updated last year
- Official implementation of Stochastic Taylor Derivative Estimator (STDE) NeurIPS2024☆115Updated 8 months ago
- ☆13Updated 6 months ago
- ☆32Updated 10 months ago
- Parallelizing non-linear sequential models over the sequence length☆53Updated last month
- Pytorch implementation of KFAC - this is a port of https://github.com/tensorflow/kfac/☆26Updated last year
- DoG is SGD's Best Friend: A Parameter-Free Dynamic Step Size Schedule☆63Updated last year
- 😎 A curated list of tensor decomposition resources for model compression.☆77Updated last week
- SLTrain: a sparse plus low-rank approach for parameter and memory efficient pretraining (NeurIPS 2024)☆32Updated 9 months ago
- [ICML 2024 Oral] LSH-Based Efficient Point Transformer (HEPT)☆20Updated 6 months ago
- LongLLaDA: Unlocking Long Context Capabilities in Diffusion LLMs☆26Updated 2 weeks ago
- Neural Tangent Kernel Papers☆115Updated 6 months ago
- ☆19Updated 4 months ago
- Pytorch code for experiments on Linear Transformers☆21Updated last year
- Tensor-Train decomposition in pytorch☆72Updated 6 months ago
- AdaSplash: Adaptive Sparse Flash Attention (aka Flash Entmax Attention)☆19Updated 3 weeks ago
- [ICLR'24] "DeepZero: Scaling up Zeroth-Order Optimization for Deep Model Training" by Aochuan Chen*, Yimeng Zhang*, Jinghan Jia, James Di…☆66Updated 10 months ago
- source code for paper "Riemannian Preconditioned LoRA for Fine-Tuning Foundation Models"☆30Updated last year
- Distributed K-FAC preconditioner for PyTorch☆89Updated this week
- ☆34Updated 4 months ago
- [NAACL 24 Oral] LoRETTA: Low-Rank Economic Tensor-Train Adaptation for Ultra-Low-Parameter Fine-Tuning of Large Language Models☆36Updated 7 months ago
- Welcome to the 'In Context Learning Theory' Reading Group☆29Updated 9 months ago
- summer school materials☆44Updated 2 years ago
- ☆81Updated last year