NVIDIA / RTX-AI-ToolkitLinks
The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PCs and cloud.
☆180Updated last month
Alternatives and similar repositories for RTX-AI-Toolkit
Users that are interested in RTX-AI-Toolkit are comparing it to the libraries listed below
Sorting:
- An NVIDIA AI Workbench example project for customizing an SDXL model☆58Updated 2 months ago
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆127Updated last year
- An NVIDIA AI Workbench example project for fine-tuning a Nemotron-3 8B model☆62Updated last year
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆198Updated 8 months ago
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆213Updated last month
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆69Updated last year
- An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)☆361Updated 4 months ago
- ☆179Updated this week
- llama.cpp fork used by GPT4All☆55Updated 10 months ago
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆72Updated last year
- ☆94Updated last year
- automatically quant GGUF models☆219Updated 2 weeks ago
- Blueprint for Ingesting massive volumes of live or archived videos and extract insights for summarization and interactive Q&A☆368Updated last month
- A sleek and user-friendly interface for interacting with Ollama models, built with Python and Gradio.☆35Updated 8 months ago
- ☆119Updated last month
- Voice Agent Framework for Conversational AI☆71Updated 8 months ago
- The NVIDIA NeMo Agent Toolkit UI streamlines interacting with NeMo Agent Toolkit workflows in an easy-to-use web application.☆63Updated this week
- Examples of RAG using Llamaindex with local LLMs - Gemma, Mixtral 8x7B, Llama 2, Mistral 7B, Orca 2, Phi-2, Neural 7B☆131Updated last year
- Simple UI for Llama-3.2-11B-Vision & Molmo-7B-D☆135Updated last year
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Using the moondream VLM with optical flow for promptable object tracking☆72Updated 10 months ago
- Unsloth Studio☆122Updated 9 months ago
- No-code CLI designed for accelerating ONNX workflows☆221Updated 7 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆167Updated 8 months ago
- Route LLM requests to the best model for the task at hand.☆149Updated last week
- AMD related optimizations for transformer models☆96Updated 2 months ago
- Collection of step-by-step playbooks for setting up AI/ML workloads on NVIDIA DGX Spark devices with Blackwell architecture.☆318Updated this week
- ☆108Updated 4 months ago
- GRadient-INformed MoE☆264Updated last year
- ☆101Updated last year