BYU-PCCL / prompt-compression-contrastive-coding
Companion repository to "Prompt Compression and Contrastive Conditioning for Controllability and Toxicity Reduction in Language Models"
☆13Updated last year
Alternatives and similar repositories for prompt-compression-contrastive-coding:
Users that are interested in prompt-compression-contrastive-coding are comparing it to the libraries listed below
- The official repository for our paper "The Dual Form of Neural Networks Revisited: Connecting Test Time Predictions to Training Patterns …☆16Updated last year
- ☆22Updated 3 years ago
- Efficient Scaling laws and collaborative pretraining.☆14Updated 3 weeks ago
- ☆18Updated 8 months ago
- Official code for the paper "Attention as a Hypernetwork"☆24Updated 8 months ago
- Fine-Tuning Pre-trained Transformers into Decaying Fast Weights☆19Updated 2 years ago
- Code for the paper "Data Feedback Loops: Model-driven Amplification of Dataset Biases"☆15Updated 2 years ago
- ☆44Updated last year
- Repo for "Zemi: Learning Zero-Shot Semi-Parametric Language Models from Multiple Tasks" ACL 2023 Findings☆16Updated last year
- Minimum Description Length probing for neural network representations☆18Updated 3 weeks ago
- ☆26Updated last year
- Official repository of paper "RNNs Are Not Transformers (Yet): The Key Bottleneck on In-context Retrieval"☆25Updated 10 months ago
- ☆27Updated last month
- Repository for Skill Set Optimization☆12Updated 6 months ago
- [NeurIPS 2023] Sparse Modular Activation for Efficient Sequence Modeling☆35Updated last year
- Official implementation of the transformer (TF) architecture suggested in a paper entitled "Looped Transformers as Programmable Computers…☆24Updated last year
- JAX implementation of "Fine-Tuning Language Models with Just Forward Passes"☆19Updated last year
- Engineering the state of RNN language models (Mamba, RWKV, etc.)☆32Updated 8 months ago
- Official code for "Accelerating Feedforward Computation via Parallel Nonlinear Equation Solving", ICML 2021☆27Updated 3 years ago
- Implementation of Token Shift GPT - An autoregressive model that solely relies on shifting the sequence space for mixing☆48Updated 3 years ago
- ☆17Updated 2 years ago
- ☆28Updated last year
- Code for the paper "Stack Attention: Improving the Ability of Transformers to Model Hierarchical Patterns"☆17Updated 11 months ago
- ☆33Updated last year
- ☆17Updated 7 months ago
- Self-Supervised Alignment with Mutual Information☆16Updated 8 months ago
- ☆13Updated last month
- Embroid: Unsupervised Prediction Smoothing Can Improve Few-Shot Classification☆11Updated last year
- Code for paper "Do Language Models Have Beliefs? Methods for Detecting, Updating, and Visualizing Model Beliefs"☆28Updated 2 years ago
- Adding new tasks to T0 without catastrophic forgetting☆32Updated 2 years ago