StonyBrookNLP / ireneLinks
[ACL 2021] IrEne: Interpretable Energy Prediction for Transformers
☆10Updated 3 years ago
Alternatives and similar repositories for irene
Users that are interested in irene are comparing it to the libraries listed below
Sorting:
- My explorations into editing the knowledge and memories of an attention network☆35Updated 2 years ago
- QAmeleon introduces synthetic multilingual QA data using PaLM, a 540B large language model. This dataset was generated by prompt tuning P…☆34Updated last year
- Implementation of TableFormer, Robust Transformer Modeling for Table-Text Encoding, in Pytorch☆39Updated 3 years ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundry☆42Updated last year
- Code repository for the public reproduction of the language modelling experiments on "MatFormer: Nested Transformer for Elastic Inference…☆22Updated last year
- Triton Implementation of HyperAttention Algorithm☆48Updated last year
- ☆44Updated last year
- Embroid: Unsupervised Prediction Smoothing Can Improve Few-Shot Classification☆11Updated last year
- ☆54Updated 10 months ago
- Official repository for the paper "Approximating Two-Layer Feedforward Networks for Efficient Transformers"☆37Updated last year
- ☆28Updated 2 years ago
- DPO, but faster 🚀☆42Updated 6 months ago
- A dashboard for exploring timm learning rate schedulers☆19Updated 6 months ago
- AutoMoE: Neural Architecture Search for Efficient Sparsely Activated Transformers☆46Updated 2 years ago
- Open sourced backend for Martian's LLM Inference Provider Leaderboard☆18Updated 9 months ago
- Efficient Deep Learning Survey Paper☆33Updated 2 years ago
- ☆12Updated 3 years ago
- ☆9Updated last year
- Repository for CPU Kernel Generation for LLM Inference☆26Updated last year
- ☆64Updated last year
- The repository contains code for Adaptive Data Optimization☆24Updated 5 months ago
- JORA: JAX Tensor-Parallel LoRA Library (ACL 2024)☆33Updated last year
- Repository for Skill Set Optimization☆13Updated 10 months ago
- A toolkit for scaling law research ⚖☆49Updated 4 months ago
- ☆13Updated 3 weeks ago
- ☆38Updated last year
- The source code of our work "Prepacking: A Simple Method for Fast Prefilling and Increased Throughput in Large Language Models" [AISTATS …☆59Updated 7 months ago
- ☆31Updated last month
- Code for NeurIPS LLM Efficiency Challenge☆59Updated last year
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆36Updated last year