Beginning of dialog window. Escape will cancel and close the window.
End of dialog window.
详情
字幕
Maximize AI Inference Serving Performance with NVIDIA Triton Inference Server
, NVIDIA
, NVIDIA
Learn what’s new in NVIDIA Triton Inference Server. NVIDIA Triton is an open-source inference serving software that simplifies the deployment of AI models at scale in production. Deploy deep learning and machine learning models from any framework (TensorFlow, NVIDIA TensorRT, PyTorch, OpenVINO, ONNX Runtime, XGBoost, or custom) on any GPU- or CPU-based infrastructure with Triton. We'll discuss some of the new backends, support for embedded devices, new integrations on the public cloud, model ensembles, and other new features. We'll also look at new customer use cases. Learn how to use Triton easily in your AI workflows and maximize the AI performance from your GPU and CPU servers across cloud, data center, and edge.