NVIDIA / RTX-AI-ToolkitLinks
The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PCs and cloud.
☆181Updated 2 months ago
Alternatives and similar repositories for RTX-AI-Toolkit
Users that are interested in RTX-AI-Toolkit are comparing it to the libraries listed below
Sorting:
- An NVIDIA AI Workbench example project for customizing an SDXL model☆60Updated 2 months ago
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆127Updated last year
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆200Updated 8 months ago
- An NVIDIA AI Workbench example project for fine-tuning a Nemotron-3 8B model☆65Updated last year
- ☆182Updated 2 weeks ago
- An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)☆362Updated 5 months ago
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆218Updated 2 weeks ago
- Simple UI for Llama-3.2-11B-Vision & Molmo-7B-D☆134Updated last year
- llama.cpp fork used by GPT4All☆55Updated 11 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆42Updated last year
- Optimized local inference for LLMs with HuggingFace-like APIs for quantization, vision/language models, multimodal agents, speech, vector…☆350Updated last year
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆170Updated 9 months ago
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆69Updated last year
- The NVIDIA NeMo Agent Toolkit UI streamlines interacting with NeMo Agent Toolkit workflows in an easy-to-use web application.☆72Updated this week
- automatically quant GGUF models☆219Updated last month
- Unsloth Studio☆125Updated 9 months ago
- Examples of RAG using Llamaindex with local LLMs - Gemma, Mixtral 8x7B, Llama 2, Mistral 7B, Orca 2, Phi-2, Neural 7B☆130Updated last year
- GPT-4 Level Conversational QA Trained In a Few Hours☆65Updated last year
- ☆123Updated 2 months ago
- Running Microsoft's BitNet via Electron, React & Astro☆51Updated 4 months ago
- Collection of reference workflows for building intelligent agents with NIMs☆187Updated last year
- Tcurtsni: Reverse Instruction Chat, ever wonder what your LLM wants to ask you?☆23Updated last year
- Blueprint for Ingesting massive volumes of live or archived videos and extract insights for summarization and interactive Q&A☆387Updated last month
- ☆101Updated last year
- Context-Aware RAG library for Knowledge Graph ingestion and retrieval functions.☆54Updated this week
- ☆120Updated 5 months ago
- Utils for Unsloth https://github.com/unslothai/unsloth☆187Updated last week
- An NVIDIA AI Workbench example project for an Agentic Retrieval Augmented Generation (RAG)☆147Updated 2 weeks ago
- An open source implementation of LFMs from Liquid AI: Liquid Foundation Models☆119Updated last year
- Run LLMs in the Browser with MLC / WebLLM ✨☆149Updated last year