foundation-model-stack / fms-hf-tuningLinks
π Collection of tuning recipes with HuggingFace SFTTrainer and PyTorch FSDP.
β47Updated this week
Alternatives and similar repositories for fms-hf-tuning
Users that are interested in fms-hf-tuning are comparing it to the libraries listed below
Sorting:
- IBM development fork of https://github.com/huggingface/text-generation-inferenceβ61Updated 2 months ago
- Python library for Synthetic Data Generationβ42Updated this week
- LM engine is a library for pretraining/finetuning LLMsβ59Updated this week
- π¦ Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data β¦β206Updated this week
- β40Updated last year
- Estimate resources needed to train LLMsβ13Updated 4 months ago
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data generaβ¦β73Updated 2 weeks ago
- β52Updated 8 months ago
- Train, tune, and infer Bamba modelβ130Updated last month
- β30Updated 8 months ago
- Google TPU optimizations for transformers modelsβ116Updated 5 months ago
- β23Updated 2 years ago
- Codebase accompanying the Summary of a Haystack paper.β79Updated 9 months ago
- Pre-training code for CrystalCoder 7B LLMβ54Updated last year
- QAlign is a new test-time alignment approach that improves language model performance by using Markov chain Monte Carlo methods.β23Updated 3 months ago
- Source code for the collaborative reasoner research project at Meta FAIR.β95Updated 3 months ago
- Supervised instruction finetuning for LLM with HF trainer and Deepspeedβ35Updated 2 years ago
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Dataβ43Updated this week
- Supercharge huggingface transformers with model parallelism.β77Updated 9 months ago
- π Collection of libraries used with fms-hf-tuning to accelerate fine-tuning and training of large models.β11Updated last month
- Open Implementations of LLM Analysesβ105Updated 9 months ago
- Accelerating your LLM training to full speed! Made with β€οΈ by ServiceNow Researchβ211Updated this week
- Repo hosting codes and materials related to speeding LLMs' inference using token merging.β36Updated last year
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing Systemβ128Updated last year
- codebase release for EMNLP2023 paper publicationβ19Updated 2 months ago
- β41Updated 5 months ago
- β48Updated 5 months ago
- Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in Alignmentβ60Updated 10 months ago
- A framework for standardizing evaluations of large foundation models, beyond single-score reporting and rankings.β163Updated this week
- Verifiers for LLM Reinforcement Learningβ65Updated 3 months ago