chujiezheng / chat_templates
Chat Templates for π€ HuggingFace Large Language Models
β588Updated last month
Alternatives and similar repositories for chat_templates:
Users that are interested in chat_templates are comparing it to the libraries listed below
- Official repository for "Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing". Your efficient and high-quality sβ¦β565Updated last week
- Evaluate your LLM's response with Prometheus and GPT4 π―β841Updated last week
- β484Updated last month
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β687Updated 3 months ago
- Official repository for ORPOβ430Updated 7 months ago
- β812Updated last week
- An Open Source Toolkit For LLM Distillationβ425Updated last week
- Implementation of paper Data Engineering for Scaling Language Models to 128K Contextβ447Updated 9 months ago
- Official Implementation of EAGLE-1 (ICML'24) and EAGLE-2 (EMNLP'24)β912Updated 2 weeks ago
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuningβ634Updated 7 months ago
- Generative Representational Instruction Tuningβ584Updated 2 months ago
- A library for easily merging multiple LLM experts, and efficiently train the merged LLM.β426Updated 4 months ago
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backendsβ970Updated this week
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.β538Updated 10 months ago
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Rewardβ800Updated 2 months ago
- [NeurIPS'24 Spotlight] To speed up Long-context LLMs' inference, approximate and dynamic sparse calculate the attention, which reduces inβ¦β874Updated 3 weeks ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ346Updated 4 months ago
- Codebase for Merging Language Models (ICML 2024)β792Updated 8 months ago
- β303Updated 7 months ago
- [ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruningβ578Updated 10 months ago
- Codes for the paper "βBench: Extending Long Context Evaluation Beyond 100K Tokens": https://arxiv.org/abs/2402.13718β303Updated 3 months ago
- Code for Quiet-STaRβ698Updated 4 months ago
- RewardBench: the first evaluation tool for reward models.β491Updated last week
- Minimalistic large language model 3D-parallelism trainingβ1,386Updated this week
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifiβ¦β1,879Updated this week
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).β785Updated 2 weeks ago
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]β524Updated last month
- awesome synthetic (text) datasetsβ253Updated 2 months ago
- LongBench v2 and LongBench (ACL 2024)β755Updated this week
- YaRN: Efficient Context Window Extension of Large Language Modelsβ1,398Updated 9 months ago