huggingface / optimum-graphcoreLinks
Blazing fast training of π€ Transformers on Graphcore IPUs
β85Updated last year
Alternatives and similar repositories for optimum-graphcore
Users that are interested in optimum-graphcore are comparing it to the libraries listed below
Sorting:
- β67Updated 2 years ago
- β186Updated last week
- Implementation of a Transformer, but completely in Tritonβ265Updated 3 years ago
- Implementation of the specific Transformer architecture from PaLM - Scaling Language Modeling with Pathways - in Jax (Equinox framework)β187Updated 2 years ago
- Training material for IPU users: tutorials, feature examples, simple applicationsβ86Updated 2 years ago
- Inference code for LLaMA models in JAXβ117Updated last year
- Load compute kernels from the Hubβ139Updated this week
- JAX implementation of the Llama 2 modelβ217Updated last year
- β59Updated 3 years ago
- Various transformers for FSDP researchβ37Updated 2 years ago
- This repository contains the experimental PyTorch native float8 training UXβ223Updated 10 months ago
- Train very large language models in Jax.β204Updated last year
- Easy and lightning fast training of π€ Transformers on Habana Gaudi processor (HPU)β186Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training easy, efficient, and effective.β166Updated this week
- Google TPU optimizations for transformers modelsβ112Updated 4 months ago
- β250Updated 10 months ago
- [WIP] A π₯ interface for running code in the cloudβ85Updated 2 years ago
- Experiment of using Tangent to autodiff tritonβ79Updated last year
- β108Updated last year
- A performant, memory-efficient checkpointing library for PyTorch applications, designed with large, complex distributed workloads in mindβ¦β157Updated 5 months ago
- Exploring finetuning public checkpoints on filter 8K sequences on Pileβ114Updated 2 years ago
- Large scale 4D parallelism pre-training for π€ transformers in Mixture of Experts *(still work in progress)*β82Updated last year
- jax-triton contains integrations between JAX and OpenAI Tritonβ395Updated this week
- Amos optimizer with JEstimator lib.β82Updated last year
- OSLO: Open Source for Large-scale Optimizationβ174Updated last year
- Pipeline for pulling and processing online language model pretraining data from the webβ178Updated last year
- Torch Distributed Experimentalβ117Updated 9 months ago
- Accelerated inference of π€ models using FuriosaAI NPU chips.β26Updated 11 months ago
- Applied AI experiments and examples for PyTorchβ271Updated this week
- Techniques used to run BLOOM at inference in parallelβ37Updated 2 years ago