Video Player is loading.
Current Time 0:00
Duration 0:00
Loaded: 0%
Stream Type LIVE
Remaining Time 0:00
 
1x
    • Chapters
    • descriptions off, selected
    • subtitles off, selected
      • Quality

      Maximize AI Inference Serving Performance with NVIDIA Triton Inference Server

      , NVIDIA
      , NVIDIA
      Learn what’s new in NVIDIA Triton Inference Server. NVIDIA Triton is an open-source inference serving software that simplifies the deployment of AI models at scale in production. Deploy deep learning and machine learning models from any framework (TensorFlow, NVIDIA TensorRT, PyTorch, OpenVINO, ONNX Runtime, XGBoost, or custom) on any GPU- or CPU-based infrastructure with Triton. We'll discuss some of the new backends, support for embedded devices, new integrations on the public cloud, model ensembles, and other new features. We'll also look at new customer use cases. Learn how to use Triton easily in your AI workflows and maximize the AI performance from your GPU and CPU servers across cloud, data center, and edge.
      活动: GTC Digital November
      日期: November 2021
      行业: All Industries
      级别: Beginner Technical
      话题: Deep Learning - Inference
      语言: English
      话题: Deep Learning - Frameworks
      所在地: