foundation-model-stack / fms-acceleration
π Collection of libraries used with fms-hf-tuning to accelerate fine-tuning and training of large models.
β9Updated last week
Alternatives and similar repositories for fms-acceleration:
Users that are interested in fms-acceleration are comparing it to the libraries listed below
- β10Updated last month
- NAACL '24 (Best Demo Paper RunnerUp) / MlSys @ NeurIPS '23 - RedCoast: A Lightweight Tool to Automate Distributed Training and Inferenceβ64Updated 4 months ago
- The simplest, fastest repository for training/finetuning medium-sized GPTs.β18Updated 2 years ago
- code for paper "Accessing higher dimensions for unsupervised word translation"β21Updated last year
- Source-to-Source Debuggable Derivatives in Pure Pythonβ15Updated last year
- Using FlexAttention to compute attention with different masking patternsβ43Updated 6 months ago
- Demo of the unit_scaling library, showing how a model can be easily adapted to train in FP8.β45Updated 8 months ago
- Minimum Description Length probing for neural network representationsβ19Updated 2 months ago
- Aioli: A unified optimization framework for language model data mixingβ23Updated 2 months ago
- PostText is a QA system for querying your text data. When appropriate structured views are in place, PostText is good at answering querieβ¦β31Updated last year
- Experimental scripts for researching data adaptive learning rate scheduling.β23Updated last year
- β21Updated last month
- Compression for Foundation Modelsβ31Updated 2 weeks ago
- β9Updated last year
- Hacks for PyTorchβ19Updated last year
- Simplifying parsing of large jsonline files in NLP Workflowsβ12Updated 3 years ago
- β18Updated 11 months ago
- Code for paper: "Privately generating tabular data using language models".β15Updated last year
- Personal solutions to the Triton Puzzlesβ18Updated 8 months ago
- Utilities for Training Very Large Modelsβ58Updated 6 months ago
- PyTorch centric eager mode debuggerβ46Updated 3 months ago
- Repository for Sparse Finetuning of LLMs via modified version of the MosaicML llmfoundryβ40Updated last year
- Train, tune, and infer Bamba modelβ88Updated 2 months ago
- Make triton easierβ47Updated 10 months ago
- Training hybrid models for dummies.β20Updated 2 months ago
- β25Updated last year
- Dolomite Engine is a library for pretraining/finetuning LLMsβ46Updated last week
- This is a new metric that can be used to evaluate faithfulness of text generated by LLMs. The work behind this repository can be found heβ¦β31Updated last year
- Efficient and Scalable Estimation of Tool Representations in Vector Spaceβ23Updated 7 months ago
- Here we will test various linear attention designs.β60Updated 11 months ago