Your Personalized AI Chatbot
System Requirements
Chat with RTX is a demo app that lets you personalize a GPT large language model (LLM) connected to your own content—docs, notes, or other data. Leveraging retrieval-augmented generation (RAG), TensorRT-LLM, and RTX acceleration, you can query a custom chatbot to quickly get contextually relevant answers. And because it all runs locally on your Windows RTX PC or workstation, you’ll get fast and secure results.
Simply download, install, and start chatting right away.
Chat with RTX supports various file formats, including text, pdf, doc/docx, and xml. Simply point the application at the folder containing your files and it'll load them into the library in a matter of seconds.
The Chat with RTX tech demo is built from the TensorRT-LLM RAG developer reference project available from GitHub. Developers can use that reference to develop and deploy their own RAG-based applications for RTX, accelerated by TensorRT-LLM.
Get next-level AI performance on GeForce RTX™ and NVIDIA RTX™ GPUs. From enhanced creativity and productivity to blisteringly fast gaming, the ultimate in AI power on Windows PCs is on RTX.
Explore NVIDIA’s generative AI developer tools and enterprise solutions.
NVIDIA Privacy Policy