NVIDIA / nim-anywhere
Accelerate your Gen AI with NVIDIA NIM and NVIDIA AI Workbench
☆144Updated 2 weeks ago
Alternatives and similar repositories for nim-anywhere:
Users that are interested in nim-anywhere are comparing it to the libraries listed below
- A collection of YAML files, Helm Charts, Operator code, and guides to act as an example reference implementation for NVIDIA NIM deploymen…☆156Updated this week
- An NVIDIA AI Workbench example project for Retrieval Augmented Generation (RAG)☆302Updated 2 months ago
- ☆132Updated this week
- Self-host LLMs with vLLM and BentoML☆86Updated this week
- ☆157Updated last week
- Collection of reference workflows for building intelligent agents with NIMs☆145Updated 3 weeks ago
- An NVIDIA AI Workbench example project for fine-tuning a Mistral 7B model☆52Updated 8 months ago
- An NVIDIA AI Workbench example project for an Agentic Retrieval Augmented Generation (RAG)☆55Updated last week
- End-to-End LLM Guide☆101Updated 7 months ago
- 🚀 Use NVIDIA NIMs with Haystack pipelines☆30Updated 5 months ago
- GenAI components at micro-service level; GenAI service composer to create mega-service☆102Updated this week
- Customizable, AI-driven virtual assistant designed to streamline customer service operations, handle common inquiries, and improve overal…☆82Updated 3 weeks ago
- Hugging Face Deep Learning Containers (DLCs) for Google Cloud☆141Updated 2 weeks ago
- ☆52Updated last month
- This reference can be used with any existing OpenAI integrated apps to run with TRT-LLM inference locally on GeForce GPU on Windows inste…☆119Updated 11 months ago
- ☆156Updated this week
- Python SDK for Llama Stack☆128Updated this week
- NVIDIA AI Blueprint for digital human for customer service.☆216Updated last week
- The NVIDIA RTX™ AI Toolkit is a suite of tools and SDKs for Windows developers to customize, optimize, and deploy AI models across RTX PC…☆138Updated 2 months ago
- Triton CLI is an open source command line interface that enables users to create, deploy, and profile models served by the Triton Inferen…☆53Updated this week
- Easy and lightning fast training of 🤗 Transformers on Habana Gaudi processor (HPU)☆169Updated this week
- Tutorial for building LLM router☆179Updated 6 months ago
- ☆223Updated this week
- ☆251Updated 2 months ago
- Evaluate and Enhance Your LLM Deployments for Real-World Inference Needs☆190Updated this week
- Build an agentic RAG system with Granite 3.1 on your laptop☆86Updated 3 weeks ago
- A Lightweight Library for AI Observability☆232Updated last week
- Pretrain, finetune and serve LLMs on Intel platforms with Ray☆110Updated this week