Chat with RTX is a demo app that lets you personalize a GPT large language model (LLM) connected to your own content—docs, notes, or other data. Leveraging retrieval-augmented generation (RAG), TensorRT-LLM, and RTX acceleration, you can query a custom chatbot to quickly get contextually relevant answers. And because it all runs locally on your Windows RTX PC or workstation, you’ll get fast and secure results.
Simply download, install, and start chatting right away.
Platform | Windows |
GPU | NVIDIA GeForce™ RTX 30 or 40 Series GPU or NVIDIA RTX™ Ampere or Ada Generation GPU with at least 8GB of VRAM |
RAM | 16GB or greater |
OS | Windows 11 |
Driver | 535.11 or later |
File Size | 35 GB |
Chat with RTX supports various file formats, including text, pdf, doc/docx, and xml. Simply point the application at the folder containing your files and it'll load them into the library in a matter of seconds.
The Chat with RTX tech demo is built from the TensorRT-LLM RAG developer reference project available from GitHub. Developers can use that reference to develop and deploy their own RAG-based applications for RTX, accelerated by TensorRT-LLM.
Get next-level AI performance on GeForce RTX™ and NVIDIA RTX™ GPUs. From enhanced creativity and productivity to blisteringly fast gaming, the ultimate in AI power on Windows PCs is on RTX.
Explore NVIDIA’s generative AI developer tools and enterprise solutions.