mosaicml / composer
Supercharge Your Model Training
☆5,116Updated this week
Related projects: ⓘ
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆7,687Updated this week
- Accessible large language models via k-bit quantization for PyTorch.☆6,029Updated this week
- Aim 💫 — An easy-to-use & supercharged open-source experiment tracker.☆5,145Updated this week
- PyTorch extensions for high performance and large scale training.☆3,149Updated 2 weeks ago
- 🚀 Accelerate training and inference of 🤗 Transformers and 🤗 Diffusers with easy to use hardware optimization tools☆2,459Updated this week
- Foundation Architecture for (M)LLMs☆3,003Updated 5 months ago
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (N…☆4,531Updated this week
- Train transformer language models with reinforcement learning.☆9,288Updated this week
- Fast and memory-efficient exact attention☆13,401Updated this week
- LLM training code for Databricks foundation models☆3,964Updated this week
- Flax is a neural network library for JAX that is designed for flexibility.☆5,950Updated this week
- Sparsity-aware deep learning inference runtime for CPUs☆2,975Updated 2 months ago
- 🤗 Evaluate: A library for easily evaluating machine learning models and datasets.☆1,965Updated this week
- A repo for distributed training of language models with Reinforcement Learning via Human Feedback (RLHF)☆4,442Updated 8 months ago
- A high-performance Python-based I/O system for large (and small) deep learning problems, with strong support for PyTorch.☆2,206Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.☆8,351Updated this week
- A Data Streaming Library for Efficient Neural Network Training☆1,076Updated this week
- Transformer related optimization, including BERT, GPT☆5,773Updated 5 months ago
- Implementation of the LLaMA language model based on nanoGPT. Supports flash attention, Int8 and GPTQ 4bit quantization, LoRA and LLaMA-Ad…☆5,958Updated last week
- A collection of libraries to optimise AI model performances☆8,373Updated last month
- FFCV: Fast Forward Computer Vision (and other ML workloads!)☆2,831Updated 3 months ago
- ☆2,635Updated last week
- General technology for enabling AI capabilities w/ LLMs and MLLMs☆3,561Updated this week
- Development repository for the Triton language and compiler☆12,698Updated this week
- Ongoing research training transformer models at scale☆9,949Updated this week
- Training and serving large-scale neural networks with auto parallelization.☆3,045Updated 9 months ago
- A concise but complete full-attention transformer with a set of promising experimental features from various papers☆4,573Updated last week
- Kernl lets you run PyTorch transformer models several times faster on GPU with a single line of code, and is designed to be easily hackab…☆1,519Updated 7 months ago
- A framework for few-shot evaluation of language models.☆6,426Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMs☆9,906Updated 3 months ago