NVIDIA-NeMo / NeMoLinks
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
β16,141Updated this week
Alternatives and similar repositories for NeMo
Users that are interested in NeMo are comparing it to the libraries listed below
Sorting:
- Ongoing research training transformer models at scaleβ14,301Updated this week
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β9,307Updated this week
- End-to-End Speech Processing Toolkitβ9,603Updated this week
- Large Language Model Text Generation Inferenceβ10,664Updated last week
- Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalitiesβ21,840Updated 4 months ago
- TensorRT LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and supports state-of-the-art optimizatβ¦β12,203Updated this week
- DeepSpeed is a deep learning optimization library that makes distributed training and inference easy, efficient, and effective.β40,803Updated this week
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β20,099Updated this week
- Development repository for the Triton language and compilerβ17,668Updated this week
- Fast inference engine for Transformer modelsβ4,154Updated this week
- Neural building blocks for speaker diarization: speech activity detection, speaker change detection, overlapped speech detection, speakerβ¦β8,726Updated last week
- Fast and memory-efficient exact attentionβ20,669Updated this week
- A PyTorch-based Speech Toolkitβ10,818Updated last week
- Welcome to the Llama Cookbook! This is your go to guide for Building with Llama: Getting started with Inference, Fine-Tuning, RAG. We alsβ¦β18,047Updated 3 weeks ago
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed librariesβ7,337Updated 2 months ago
- Hackable and optimized Transformers building blocks, supporting a composable construction.β10,117Updated last week
- ONNX Runtime: cross-platform, high performance ML inferencing and training acceleratorβ18,442Updated this week
- π Accelerate inference and training of π€ Transformers, Diffusers, TIMM and Sentence Transformers with easy to use hardware optimizationβ¦β3,188Updated last week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.β31,983Updated last month
- Unsupervised text tokenizer for Neural Network-based text generation.β11,460Updated last week
- State-of-the-Art Deep Learning scripts organized by models - easy to train and deploy with reproducible accuracy and performance on enterβ¦β14,590Updated last year
- PyTorch extensions for high performance and large scale training.β3,387Updated 7 months ago
- Train transformer language models with reinforcement learning.β16,382Updated this week
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"β6,450Updated 2 weeks ago
- Accessible large language models via k-bit quantization for PyTorch.β7,767Updated this week
- Open standard for machine learning interoperabilityβ19,933Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.β10,060Updated this week
- PyTorch native post-training libraryβ5,595Updated last week
- A library for efficient similarity search and clustering of dense vectors.β38,093Updated this week
- Transformer related optimization, including BERT, GPTβ6,354Updated last year