Tomorrowdawn / top_nsigma
The official code repo and data hub of top_nsigma sampling strategy for LLMs.
☆18Updated last month
Alternatives and similar repositories for top_nsigma:
Users that are interested in top_nsigma are comparing it to the libraries listed below
- Official implementation for 'Extending LLMs’ Context Window with 100 Samples'☆76Updated 11 months ago
- Codebase for Instruction Following without Instruction Tuning☆33Updated 3 months ago
- Official repository for paper "Weak-to-Strong Extrapolation Expedites Alignment"☆71Updated 7 months ago
- ☆17Updated 8 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignment☆53Updated 4 months ago
- ☆70Updated 3 months ago
- Parameter-Efficient Sparsity Crafting From Dense to Mixture-of-Experts for Instruction Tuning on General Tasks☆136Updated 3 months ago
- Code for Paper: Harnessing Webpage Uis For Text Rich Visual Understanding☆44Updated 3 weeks ago
- A repository for research on medium sized language models.☆76Updated 7 months ago
- FuseAI Project☆76Updated 3 weeks ago
- SuperCorrect: Supervising and Correcting Language Models with Error-Driven Insights☆43Updated 2 months ago
- ☆76Updated 2 months ago
- This is the official repository of the paper "OlympicArena: Benchmarking Multi-discipline Cognitive Reasoning for Superintelligent AI"☆89Updated 3 weeks ago
- Pytorch implementation for "Compressed Context Memory For Online Language Model Interaction" (ICLR'24)☆51Updated 8 months ago
- ☆69Updated 4 months ago
- ☆53Updated 2 months ago
- [ICLR 2024] CLEX: Continuous Length Extrapolation for Large Language Models☆75Updated 9 months ago
- Data and code for our paper "Why Does the Effective Context Length of LLMs Fall Short?"☆68Updated last month
- Official repository for the paper "SwitchHead: Accelerating Transformers with Mixture-of-Experts Attention"☆96Updated 3 months ago
- This is the official repository for Inheritune.☆107Updated 3 months ago
- Code for the arXiv preprint "The Unreasonable Effectiveness of Easy Training Data"☆45Updated 11 months ago
- Script for processing OpenAI's PRM800K process supervision dataset into an Alpaca-style instruction-response format☆27Updated last year
- From GaLore to WeLore: How Low-Rank Weights Non-uniformly Emerge from Low-Rank Gradients. Ajay Jaiswal, Lu Yin, Zhenyu Zhang, Shiwei Liu,…☆42Updated 5 months ago
- Co-LLM: Learning to Decode Collaboratively with Multiple Language Models☆105Updated 8 months ago
- Minimal implementation of the Self-Play Fine-Tuning Converts Weak Language Models to Strong Language Models paper (ArXiv 20232401.01335)☆29Updated 10 months ago
- Reformatted Alignment☆113Updated 3 months ago
- Implementation of "LM-Infinite: Simple On-the-Fly Length Generalization for Large Language Models"☆42Updated last month
- ☆31Updated this week
- Exploring Model Kinship for Merging Large Language Models☆22Updated last month
- ☆133Updated 3 months ago