foundation-model-stack / fms-hf-tuningLinks
π Collection of tuning recipes with HuggingFace SFTTrainer and PyTorch FSDP.
β47Updated this week
Alternatives and similar repositories for fms-hf-tuning
Users that are interested in fms-hf-tuning are comparing it to the libraries listed below
Sorting:
- IBM development fork of https://github.com/huggingface/text-generation-inferenceβ61Updated 3 months ago
- π¦ Unitxt is a Python library for enterprise-grade evaluation of AI performance, offering the world's largest catalog of tools and data β¦β206Updated this week
- Python library for Synthetic Data Generationβ42Updated this week
- β41Updated last year
- Google TPU optimizations for transformers modelsβ117Updated 6 months ago
- LM engine is a library for pretraining/finetuning LLMsβ61Updated this week
- Train, tune, and infer Bamba modelβ130Updated 2 months ago
- Benchmark suite for LLMs from Fireworks.aiβ76Updated last week
- Matrix (Multi-Agent daTa geneRation Infra and eXperimentation framework) is a versatile engine for multi-agent conversational data generaβ¦β81Updated last week
- β232Updated this week
- Simple and efficient DeepSeek V3 SFT using pipeline parallel and expert parallel, with both FP8 and BF16 trainingsβ68Updated last week
- β53Updated 9 months ago
- InstructLab Training Library - Efficient Fine-Tuning with Message-Format Dataβ42Updated this week
- Source code for the collaborative reasoner research project at Meta FAIR.β99Updated 3 months ago
- Source code of "How to Correctly do Semantic Backpropagation on Language-based Agentic Systems" π€β72Updated 8 months ago
- β124Updated 9 months ago
- β129Updated 4 months ago
- QAlign is a new test-time alignment approach that improves language model performance by using Markov chain Monte Carlo methods.β23Updated 4 months ago
- β31Updated 8 months ago
- codebase release for EMNLP2023 paper publicationβ19Updated 3 months ago
- The code for the paper ROUTERBENCH: A Benchmark for Multi-LLM Routing Systemβ131Updated last year
- A Python wrapper around HuggingFace's TGI (text-generation-inference) and TEI (text-embedding-inference) servers.β33Updated 2 months ago
- Accelerating your LLM training to full speed! Made with β€οΈ by ServiceNow Researchβ218Updated this week
- vLLM adapter for a TGIS-compatible gRPC server.β34Updated this week
- Hugging Face Inference Toolkit used to serve transformers, sentence-transformers, and diffusers models.β83Updated this week
- β49Updated 5 months ago
- Code for NeurIPS LLM Efficiency Challengeβ59Updated last year
- Manage scalable open LLM inference endpoints in Slurm clustersβ268Updated last year
- Utils for Unslothβ120Updated last week
- An introduction to LLM Samplingβ79Updated 7 months ago