XuehaiPan / nvitop
An interactive NVIDIA-GPU process viewer and beyond, the one-stop solution for GPU process management.
β4,617Updated last week
Related projects: β
- π A simple way to launch, train, and use PyTorch models on almost any device and distributed configuration, automatic mixed precision (iβ¦β7,687Updated this week
- Fast and memory-efficient exact attentionβ13,401Updated this week
- Hackable and optimized Transformers building blocks, supporting a composable construction.β8,351Updated this week
- Accessible large language models via k-bit quantization for PyTorch.β6,029Updated this week
- Code for loralib, an implementation of "LoRA: Low-Rank Adaptation of Large Language Models"β10,327Updated last month
- An open source implementation of CLIP.β9,782Updated last month
- π€ PEFT: State-of-the-art Parameter-Efficient Fine-Tuning.β15,839Updated this week
- Ongoing research training transformer models at scaleβ9,949Updated this week
- Flexible and powerful tensor operations for readable and reliable code (for pytorch, jax, TF and others)β8,362Updated this week
- RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the bestβ¦β12,397Updated 2 weeks ago
- Transformer related optimization, including BERT, GPTβ5,773Updated 5 months ago
- Development repository for the Triton language and compilerβ12,698Updated this week
- Running large language models on a single GPU for throughput-oriented scenarios.β9,133Updated last month
- GPU & Accelerator process monitoring for AMD, Apple, Huawei, Intel, NVIDIA and Qualcommβ7,981Updated 2 weeks ago
- π A simple command-line utility for querying and monitoring GPU statusβ4,009Updated last month
- LAVIS - A One-stop Library for Language-Vision Intelligenceβ9,663Updated 3 weeks ago
- Simple, safe way to store and distribute tensorsβ2,755Updated 2 weeks ago
- A collection of libraries to optimise AI model performancesβ8,373Updated last month
- Mamba SSM architectureβ12,542Updated last month
- A concise but complete full-attention transformer with a set of promising experimental features from various papersβ4,573Updated last week
- PyTorch code and models for the DINOv2 self-supervised learning method.β8,791Updated last month
- ImageBind One Embedding Space to Bind Them Allβ8,221Updated last month
- AITemplate is a Python framework which renders neural network into high performance CUDA/HIP C++ code. Specialized for FP16 TensorCore (Nβ¦β4,531Updated this week
- π¦ Lion, new optimizer discovered by Google Brain using genetic algorithms that is purportedly better than Adam(w), in Pytorchβ2,002Updated 3 months ago
- PyTorch extensions for high performance and large scale training.β3,149Updated 2 weeks ago
- TensorRT-LLM provides users with an easy-to-use Python API to define Large Language Models (LLMs) and build TensorRT engines that containβ¦β8,186Updated last week
- A high-throughput and memory-efficient inference and serving engine for LLMsβ26,822Updated this week
- QLoRA: Efficient Finetuning of Quantized LLMsβ9,906Updated 3 months ago
- Train transformer language models with reinforcement learning.β9,288Updated this week
- Official repo for consistency models.β6,073Updated 5 months ago