NVIDIA / RTX-AI-ToolkitLinks
The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PCs and cloud.
☆180Updated 2 weeks ago
Alternatives and similar repositories for RTX-AI-Toolkit
Users that are interested in RTX-AI-Toolkit are comparing it to the libraries listed below
Sorting:
- An NVIDIA AI Workbench example project for customizing an SDXL model☆58Updated last month
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆126Updated last year
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆192Updated 7 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)☆349Updated 3 months ago
- ☆176Updated last week
- automatically quant GGUF models☆219Updated last month
- Simple UI for Llama-3.2-11B-Vision & Molmo-7B-D☆136Updated last year
- llama.cpp fork used by GPT4All☆55Updated 9 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆163Updated 7 months ago
- Unsloth Studio☆118Updated 8 months ago
- ☆108Updated 3 months ago
- An NVIDIA AI Workbench example project for fine-tuning a Nemotron-3 8B model☆59Updated last year
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆66Updated last year
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆211Updated this week
- A Windows tool to query various LLM AIs. Supports branched conversations, history and summaries among others.☆33Updated last month
- Running Microsoft's BitNet via Electron, React & Astro☆48Updated 2 months ago
- ☆94Updated 11 months ago
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆69Updated last year
- ☆266Updated 5 months ago
- LLM inference in C/C++☆103Updated this week
- Gradio based tool to run opensource LLM models directly from Huggingface☆96Updated last year
- Utils for Unsloth https://github.com/unslothai/unsloth☆180Updated this week
- GRadient-INformed MoE☆264Updated last year
- Self-host LLMs with vLLM and BentoML☆161Updated 2 weeks ago
- Lightweight continuous batching OpenAI compatibility using HuggingFace Transformers include T5 and Whisper.☆29Updated 8 months ago
- AMD related optimizations for transformer models☆96Updated last month
- Examples of RAG using Llamaindex with local LLMs - Gemma, Mixtral 8x7B, Llama 2, Mistral 7B, Orca 2, Phi-2, Neural 7B☆132Updated last year
- ☆126Updated last year
- Gemma 2 optimized for your local machine.☆378Updated last year