PKU-ML / LongPPL
☆15Updated last week
Related projects ⓘ
Alternatives and complementary repositories for LongPPL
- Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]☆14Updated 6 months ago
- Long Context Extension and Generalization in LLMs☆39Updated last month
- [ATTRIB @ NeurIPS 2024] When Attention Sink Emerges in Language Models: An Empirical View☆27Updated 3 weeks ago
- Codebase for decoding compressed trust.☆20Updated 6 months ago
- [NeurIPS-2024] 📈 Scaling Laws with Vocabulary: Larger Models Deserve Larger Vocabularies https://arxiv.org/abs/2407.13623☆67Updated last month
- The code of “Improving Weak-to-Strong Generalization with Scalable Oversight and Ensemble Learning”☆15Updated 8 months ago
- Source code for the TMLR paper "Black-Box Prompt Learning for Pre-trained Language Models"☆55Updated last year
- ☆19Updated last month
- [ICML 2024] Junk DNA Hypothesis: A Task-Centric Angle of LLM Pre-trained Weights through Sparsity; Lu Yin*, Ajay Jaiswal*, Shiwei Liu, So…☆15Updated 5 months ago
- ☆44Updated 10 months ago
- Official Repository for The Paper: Safety Alignment Should Be Made More Than Just a Few Tokens Deep☆26Updated 4 months ago
- ☆33Updated last year
- ☆26Updated 6 months ago
- A curated list of awesome resources dedicated to Scaling Laws for LLMs☆63Updated last year
- Code for paper "Diffusion Language Models Can Perform Many Tasks with Scaling and Instruction-Finetuning"☆63Updated 9 months ago
- [ICLR 2023] "Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable Transformers" by Tianlong Chen*, Zhenyu Zhang*, Ajay Jaiswal…☆44Updated last year
- Repository for "Propagating Knowledge Updates to LMs Through Distillation" (NeurIPS 2023).☆24Updated 2 months ago
- Test-time-training on nearest neighbors for large language models☆25Updated 6 months ago
- ☆31Updated last year
- Learning adapter weights from task descriptions☆15Updated last year
- [ICLR'24] RAIN: Your Language Models Can Align Themselves without Finetuning☆83Updated 5 months ago
- `dattri` is a PyTorch library for developing, benchmarking, and deploying efficient data attribution algorithms.☆31Updated last week
- Official repository of "Localizing Task Information for Improved Model Merging and Compression" [ICML 2024]☆34Updated 2 weeks ago
- Stick-breaking attention☆33Updated this week
- Restore safety in fine-tuned language models through task arithmetic☆26Updated 7 months ago
- [NeurIPS'23] Aging with GRACE: Lifelong Model Editing with Discrete Key-Value Adaptors☆69Updated 8 months ago
- [NeurIPS 2024] Knowledge Circuits in Pretrained Transformers☆68Updated 3 weeks ago
- Lightweight Adapting for Black-Box Large Language Models☆18Updated 8 months ago
- ☆43Updated 9 months ago
- Methods and evaluation for aligning language models temporally☆24Updated 8 months ago