philschmid / knowledge-distillation-transformers-pytorch-sagemakerLinks
☆47Updated 3 years ago
Alternatives and similar repositories for knowledge-distillation-transformers-pytorch-sagemaker
Users that are interested in knowledge-distillation-transformers-pytorch-sagemaker are comparing it to the libraries listed below
Sorting:
- Scalable training for dense retrieval models.☆299Updated last month
- DSIR large-scale data selection framework for language model training☆253Updated last year
- Code and data for "Lost in the Middle: How Language Models Use Long Contexts"☆351Updated last year
- Finetune mistral-7b-instruct for sentence embeddings☆85Updated last year
- Official repository for ORPO☆459Updated last year
- A framework for few-shot evaluation of autoregressive language models.☆105Updated 2 years ago
- Benchmarking library for RAG☆216Updated last week
- Multilingual Large Language Models Evaluation Benchmark☆127Updated 11 months ago
- Official repository of NEFTune: Noisy Embeddings Improves Instruction Finetuning☆397Updated last year
- ☆284Updated last year
- [EMNLP 2023] The CoT Collection: Improving Zero-shot and Few-shot Learning of Language Models via Chain-of-Thought Fine-Tuning☆245Updated last year
- Implementation of paper Data Engineering for Scaling Language Models to 128K Context☆467Updated last year
- Code for Multilingual Eval of Generative AI paper published at EMNLP 2023☆70Updated last year
- MultilingualSIFT: Multilingual Supervised Instruction Fine-tuning☆94Updated last year
- [EMNLP 2023] Adapting Language Models to Compress Long Contexts☆308Updated 10 months ago
- Tk-Instruct is a Transformer model that is tuned to solve many NLP tasks by following instructions.☆180Updated 2 years ago
- AIR-Bench: Automated Heterogeneous Information Retrieval Benchmark☆150Updated 7 months ago
- What's In My Big Data (WIMBD) - a toolkit for analyzing large text datasets☆222Updated 8 months ago
- Scaling Data-Constrained Language Models☆338Updated 3 weeks ago
- ToolQA, a new dataset to evaluate the capabilities of LLMs in answering challenging questions with external tools. It offers two levels …☆272Updated last year
- This is the repository for our paper "INTERS: Unlocking the Power of Large Language Models in Search with Instruction Tuning"☆204Updated 7 months ago
- Manage scalable open LLM inference endpoints in Slurm clusters☆267Updated last year
- Reverse Instructions to generate instruction tuning data with corpus examples☆214Updated last year
- ☆524Updated 8 months ago
- Train Llama 2 & 3 on the SQuAD v2 task as an example of how to specialize a generalized (foundation) model.☆52Updated last year
- Github repository for "RAGTruth: A Hallucination Corpus for Developing Trustworthy Retrieval-Augmented Language Models"☆192Updated 7 months ago
- Official PyTorch implementation of DistiLLM: Towards Streamlined Distillation for Large Language Models (ICML 2024)☆225Updated 4 months ago
- The official evaluation suite and dynamic data release for MixEval.☆242Updated 8 months ago
- Lightweight demos for finetuning LLMs. Powered by 🤗 transformers and open-source datasets.☆77Updated 9 months ago
- Scripts for fine-tuning Llama2 via SFT and DPO.☆200Updated last year