Explore the latest network fabric innovations for AI model training, inference, and retrieval-augmented generation.
As AI continues to evolve, the need for massive computational power has led to developing some of the world's largest GPU-based data centers. These centers are at the forefront of training large language models (LLMs) with trillions of parameters, pushing the boundaries of what AI can achieve. In this session, leading AI cloud data centers will come together to share our experiences and insights from building and deploying these colossal systems. We'll delve into the unique challenges of networking at such a massive scale, and how we overcame them. Attendees will gain a deep understanding of the lessons learned in scaling infrastructure to support the next generation of AI, from the complexities of connecting thousands of GPUs to the innovations required to maintain performance and reliability at such an unprecedented scale.
Important: Near capacity, highly suggest arriving early. Attendees are let in on a first-come, first-served basis.
We'll introduce newly announced networking solutions and advanced features that are specifically designed for AI data centers. By showcasing cutting-edge technologies and recent innovations, we'll explore how these developments are set to revolutionize the landscape of AI model training and deployment. These advancements will unlock new levels of performance, driving the next wave of AI advancements and enabling faster, more efficient processing of AI workloads.
Important: Near capacity, highly suggest arriving early. Attendees are let in on a first-come, first-served basis.
In the era of generative AI, retrieval augmented generation (RAG) is emerging as a vital tool, enabling enterprises to capitalize on their data. As data evolves and volumes grow exponentially, traditional storage solutions can no longer meet the rigorous demands of enterprise AI systems. GenAI inferencing has become a rack-scale computing workload, necessitating robust, enterprise-grade infrastructure. Discover how NVIDIA is collaborating with storage trailblazers to create a new class of accelerated data platforms — optimized for GenAI workloads and real-time RAG pipelines. Powering these next-generation systems are NVIDIA GPUs, high-performance NVIDIA networking, NVIDIA NIMs, and an AI-native data ingestion stack, delivering efficient pre-processing, storage, and retrieval at unprecedented speeds. Explore the innovations that are redefining enterprise data platforms and shaping the future of GenAI.
Important: Near capacity, highly suggest arriving early. Attendees are let in on a first-come, first-served basis.
The future of AI cloud computing demands infrastructure that is not only powerful but also agile, scalable, and efficient. In this joint session, NVIDIA and CoreWeave will share their collaborative journey of innovation — where cutting-edge networking meets real-world cloud-scale AI challenges. Get an exclusive look at NVIDIA’s full-stack networking platforms, designed to unlock AI’s full potential. CoreWeave, as an AI hyperscaler and early adopter, will reveal how they leverage the breadth and depth of NVIDIA’s networking platforms to transform cloud infrastructure. They’ll discuss achieving supercomputing scale with reliability, deep observability, and flexibility to support next-gen AI workloads. Join us to explore the breakthroughs, lessons, and strategies that enable AI platforms to scale without compromise. Gain insights into the innovations driving the industry forward and discover actionable takeaways to build infrastructure ready for tomorrow’s demands.
Important: Near capacity, highly suggest arriving early. Attendees are let in on a first-come, first-served basis.
Discuss with experts everything related to inter-GPU communication through NVLink, Infiniband, or other networks. We'll cover all communication libraries: NCCL, MPI, UCX/UCC and NVSHMEM. This is the perfect place to discuss performance benefits of GPU Direct, NVLink, Infiniband, and SHARP to accelerate your deep learning training workload or your HPC application.
Important: Connect with the Experts sessions are interactive sessions that give you a unique opportunity to meet, in either a group or one-on-one setting, with the minds behind NVIDIA’s products and research to get your questions answered. Attendees can meet with experts on a first-come, first-served basis.
Just as critical to the performance of AI model training, tail latency can negatively affect real-time AI inference, financial trading, and large-scale data analytics, where even microsecond delays can affect outcomes. The approach to zero-tail latency and accelerated networking with enhanced programmable I/O represent a paradigm shift in the performance and scalability of modern computing infrastructures. This session will open the discussion to dive into the significance of accelerated networking technologies in delivering zero-tail latency, ensuring that every data transfer meets demanding performance requirements even beyond 100k GPU scale.
Important: Connect with the Experts sessions are interactive sessions that give you a unique opportunity to meet, in either a group or one-on-one setting, with the minds behind NVIDIA’s products and research to get your questions answered. Attendees can meet with experts on a first-come, first-served basis.