Connect with the Experts: Optimize Deep Learning Inference Workloads using NVIDIA TensorRT and Deploying AI Models in Production with NVIDIA Triton Inference Server
, Product Manager, NVIDIA
, Technical Marketing Engineer, NVIDIA
, Deep Learning Solutions Architect, NVIDIA
, Product Manager, NVIDIA
, Group Product Manager, NVIDIA
, PRINCIPAL ENGINEER, NVIDIA
, TensorRT Engineering Manager, NVIDIA
In this session we focus on optimized deep learning inference and deployment of AI models in production. We'll discuss how the NVIDIA TensorRT, an SDK for high-performance deep learning inference, can deliver low latency and high throughput for deep learning applications. We'll then talk about fast and scalable deployment using NVIDIA Triton.
*IMPORTANT: Connect with the Experts sessions are interactive sessions that give you a unique opportunity to meet, in either a group or 1:1 setting, with the brilliant minds behind NVIDIA’s products and research to get your questions answered. Space is limited - first come, first served. We request that you limit your 1:1 discussion with our Experts to 5 minutes. You will have the option to ask questions in a group setting as well. We also recommend that you use a headset microphone to ensure our Experts can hear you clearly. To test your webcam (optional) and microphone settings, please visit this link.