NVIDIA / RTX-AI-ToolkitLinks
The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PCs and cloud.
☆158Updated 6 months ago
Alternatives and similar repositories for RTX-AI-Toolkit
Users that are interested in RTX-AI-Toolkit are comparing it to the libraries listed below
Sorting:
- An NVIDIA AI Workbench example project for customizing an SDXL model☆52Updated 3 weeks ago
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆121Updated last year
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆56Updated 7 months ago
- An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)☆317Updated 3 weeks ago
- llama.cpp fork used by GPT4All☆55Updated 3 months ago
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆165Updated last month
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆135Updated 3 months ago
- A comprehensive platform for managing, testing, and leveraging Ollama AI models with advanced features for customization, workflow automa…☆48Updated 2 months ago
- A curated list of OpenVINO based AI projects☆134Updated 5 months ago
- Gradio based tool to run opensource LLM models directly from Huggingface☆91Updated 11 months ago
- Blueprint for Ingesting massive volumes of live or archived videos and extract insights for summarization and interactive Q&A☆97Updated last month
- ☆156Updated last week
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆57Updated last year
- An NVIDIA AI Workbench example project for fine-tuning a Nemotron-3 8B model☆49Updated last year
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆43Updated 8 months ago
- Service for testing out the new Qwen2.5 omni model☆49Updated last month
- Serving LLMs in the HF-Transformers format via a PyFlask API☆71Updated 8 months ago
- ☆101Updated 9 months ago
- LM Studio JSON configuration file format and a collection of example config files.☆199Updated 10 months ago
- ☆90Updated 5 months ago
- Chat with Phi 3.5/3 Vision LLMs. Phi-3.5-vision is a lightweight, state-of-the-art open multimodal model built upon datasets which includ…☆33Updated 5 months ago
- automatically quant GGUF models☆181Updated this week
- Unsloth Studio☆87Updated 2 months ago
- Optimized local inference for LLMs with HuggingFace-like APIs for quantization, vision/language models, multimodal agents, speech, vector…☆271Updated 7 months ago
- An innovative library for efficient LLM inference via low-bit quantization☆348Updated 9 months ago
- ☆129Updated last month
- ☆160Updated 3 months ago
- A high-throughput and memory-efficient inference and serving engine for LLMs☆30Updated last year
- An NVIDIA AI Workbench example project for an Agentic Retrieval Augmented Generation (RAG)☆80Updated 3 weeks ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆148Updated last month