chujiezheng / chat_templatesLinks
Chat Templates for π€ HuggingFace Large Language Models
β672Updated 6 months ago
Alternatives and similar repositories for chat_templates
Users that are interested in chat_templates are comparing it to the libraries listed below
Sorting:
- [ICLR 2025] Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing. Your efficient and high-quality synthetic data β¦β713Updated 3 months ago
- [ICML'24 Spotlight] LLM Maybe LongLM: Self-Extend LLM Context Window Without Tuningβ653Updated last year
- Official repository for ORPOβ455Updated last year
- Lighteval is your all-in-one toolkit for evaluating LLMs across multiple backendsβ1,629Updated this week
- Memory optimization and training recipes to extrapolate language models' context length to 1 million tokens, with minimal hardware.β731Updated 8 months ago
- An Open Source Toolkit For LLM Distillationβ651Updated 2 weeks ago
- β938Updated 4 months ago
- A library with extensible implementations of DPO, KTO, PPO, ORPO, and other human-aware loss functions (HALOs).β857Updated 2 weeks ago
- [NeurIPS 2024] SimPO: Simple Preference Optimization with a Reference-Free Rewardβ897Updated 4 months ago
- RewardBench: the first evaluation tool for reward models.β604Updated last week
- Generative Representational Instruction Tuningβ651Updated 3 months ago
- [ICLR 2024] Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruningβ617Updated last year
- Official implementation for the paper "DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models"β498Updated 5 months ago
- This repository contains code to quantitatively evaluate instruction-tuned models such as Alpaca and Flan-T5 on held-out tasks.β546Updated last year
- β773Updated last month
- Deita: Data-Efficient Instruction Tuning for Alignment [ICLR2024]β556Updated 6 months ago
- β520Updated 7 months ago
- LongBench v2 and LongBench (ACL 25'&24')β903Updated 5 months ago
- Automatic evals for LLMsβ429Updated 2 weeks ago
- A library for easily merging multiple LLM experts, and efficiently train the merged LLM.β482Updated 9 months ago
- [ACL 2024] Progressive LLaMA with Block Expansion.β505Updated last year
- The code of our paper "InfLLM: Unveiling the Intrinsic Capacity of LLMs for Understanding Extremely Long Sequences with Training-Free Memβ¦β365Updated last year
- distributed trainer for LLMsβ577Updated last year
- Scalable toolkit for efficient model alignmentβ814Updated 3 weeks ago
- [ACL'24] Selective Reflection-Tuning: Student-Selected Data Recycling for LLM Instruction-Tuningβ357Updated 9 months ago
- Distilabel is a framework for synthetic data and AI feedback for engineers who need fast, reliable and scalable pipelines based on verifiβ¦β2,757Updated last week
- Evaluate your LLM's response with Prometheus and GPT4 π―β954Updated last month
- Codebase for Merging Language Models (ICML 2024)β832Updated last year
- The code and data for "MMLU-Pro: A More Robust and Challenging Multi-Task Language Understanding Benchmark" [NeurIPS 2024]β254Updated 3 months ago
- Official repository of NEFTune: Noisy Embeddings Improves Instruction Finetuningβ396Updated last year