NVIDIA-NeMo / NeMoLinks
A scalable generative AI framework built for researchers and developers working on Large Language Models, Multimodal, and Speech AI (Automatic Speech Recognition and Text-to-Speech)
☆15,838Updated this week
Alternatives and similar repositories for NeMo
Users that are interested in NeMo are comparing it to the libraries listed below
Sorting:
- Ongoing research training transformer models at scale☆13,755Updated last week
- Accessible large language models via k-bit quantization for PyTorch.☆7,647Updated last week
- An implementation of model parallel autoregressive transformers on GPUs, based on the Megatron and DeepSpeed libraries☆7,308Updated 2 weeks ago
- 🚀 A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (i…☆9,199Updated this week
- Development repository for the Triton language and compiler☆17,154Updated this week
- Facebook AI Research Sequence-to-Sequence Toolkit written in Python.☆31,861Updated last week
- Pretrain, finetune ANY AI model of ANY size on 1 or 10,000+ GPUs with zero code changes.☆30,222Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.☆9,989Updated last week
- Fast and memory-efficient exact attention☆19,864Updated this week
- State-of-the-Art Text Embeddings☆17,649Updated this week
- Run, manage, and scale AI workloads on any AI infrastructure. Use one system to access & manage all AI compute (Kubernetes, 17+ clouds, o…☆8,807Updated this week
- Fast inference engine for Transformer models☆4,056Updated 6 months ago
- Composable transformations of Python+NumPy programs: differentiate, vectorize, JIT to GPU/TPU, and more☆33,658Updated this week
- Trax — Deep Learning with Clear Code and Speed☆8,287Updated 2 weeks ago
- 💥 Fast State-of-the-Art Tokenizers optimized for Research and Production☆10,119Updated this week
- RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable)…☆14,009Updated this week
- The Triton Inference Server provides an optimized cloud and edge inferencing solution.☆9,879Updated this week
- A high-throughput and memory-efficient inference and serving engine for LLMs☆59,817Updated this week
- Neural building blocks for speaker diarization: speech activity detection, speaker change detection, overlapped speech detection, speaker…☆8,461Updated this week
- End-to-End Speech Processing Toolkit☆9,496Updated last week
- Flax is a neural network library for JAX that is designed for flexibility.☆6,845Updated this week
- PyTorch extensions for high performance and large scale training.☆3,380Updated 5 months ago
- A PyTorch-based Speech Toolkit☆10,536Updated this week
- Code and documentation to train Stanford's Alpaca models, and generate the data.☆30,170Updated last year
- The easiest way to serve AI apps and models - Build Model Inference APIs, Job queues, LLM apps, Multi-model pipelines, and more!☆8,119Updated this week
- 🤗 The largest hub of ready-to-use datasets for AI models with fast, easy-to-use and efficient data manipulation tools☆20,721Updated this week
- Code for the paper "Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer"☆6,435Updated 5 months ago
- Large Language Model Text Generation Inference☆10,550Updated 3 weeks ago
- Train transformer language models with reinforcement learning.☆15,818Updated this week
- Repo for external large-scale work☆6,546Updated last year