NVIDIA / RTX-AI-Toolkit
The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PCs and cloud.
☆149Updated 5 months ago
Alternatives and similar repositories for RTX-AI-Toolkit:
Users that are interested in RTX-AI-Toolkit are comparing it to the libraries listed below
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆120Updated last year
- An NVIDIA AI Workbench example project for customizing an SDXL model☆50Updated this week
- Gradio based tool to run opensource LLM models directly from Huggingface☆91Updated 10 months ago
- automatically quant GGUF models☆174Updated last week
- This is an NVIDIA AI Workbench example project that demonstrates an end-to-end model development workflow using Llamafactory.☆55Updated 6 months ago
- llama.cpp fork used by GPT4All☆55Updated 2 months ago
- Cortex.Tensorrt-LLM is a C++ inference library that can be loaded by any server at runtime. It submodules NVIDIA’s TensorRT-LLM for GPU a…☆43Updated 7 months ago
- ☆152Updated this week
- Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench☆159Updated last week
- LLM inference in C/C++☆76Updated this week
- Simple CogVLM client script☆14Updated last year
- GPT-4 Level Conversational QA Trained In a Few Hours☆61Updated 8 months ago
- Tool to download models from Huggingface Hub and convert them to GGML/GGUF for llama.cpp☆136Updated last week
- TechXplainator Tutorials☆19Updated 9 months ago
- Unsloth Studio☆80Updated last month
- Tcurtsni: Reverse Instruction Chat, ever wonder what your LLM wants to ask you?☆22Updated 10 months ago
- An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)☆313Updated this week
- A python package for developing AI applications with local LLMs.☆149Updated 4 months ago
- ☆89Updated 4 months ago
- ☆66Updated 11 months ago
- This extension enhances the capabilities of textgen-webui by integrating advanced vision models, allowing users to have contextualized co…☆53Updated 6 months ago
- An endpoint server for efficiently serving quantized open-source LLMs for code.☆55Updated last year
- ☆23Updated 7 months ago
- LLaVA server (llama.cpp).☆180Updated last year
- Chat with Phi 3.5/3 Vision LLMs. Phi-3.5-vision is a lightweight, state-of-the-art open multimodal model built upon datasets which includ…☆33Updated 4 months ago
- Educational framework exploring ergonomic, lightweight multi-agent orchestration. Modified to use local Ollama endpoint☆50Updated 6 months ago
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆128Updated 2 months ago
- ☆159Updated 3 months ago
- ☆130Updated 2 weeks ago
- Curated resources about automated GUI computer-use via LLMs. Highly opinionated, focus is on quality vs quantity.☆22Updated 5 months ago