Anshul Jindal & Martin Piercy
Your Next AI Needs 10,000 GPUs. Now What?
#1about 2 minutes
Introduction to large-scale AI infrastructure challenges
An overview of the topics to be covered, from the progress of generative AI to the compute requirements for training and inference.
#2about 4 minutes
Understanding the fundamental shift to generative AI
Generative AI creates novel content, moving beyond prediction to unlock new use cases in coding, content creation, and customer experience.
#3about 6 minutes
Using NVIDIA NIMs and blueprints to deploy models
NVIDIA Inference Microservices (NIMs) and blueprints provide pre-packaged, optimized containers to quickly deploy models for tasks like retrieval-augmented generation (RAG).
#4about 4 minutes
An overview of the AI model development lifecycle
Building a production-ready model involves a multi-stage process including data curation, distributed training, alignment, optimized inference, and implementing guardrails.
#5about 6 minutes
Understanding parallelism techniques for distributed AI training
Training massive models requires splitting them across thousands of GPUs using tensor, pipeline, and data parallelism to manage compute and communication.
#6about 2 minutes
The scale of GPU compute for training and inference
Training large models like Llama requires millions of GPU hours, while inference for a single large model can demand a full multi-GPU server.
#7about 3 minutes
Key hardware and network design for AI infrastructure
Effective multi-node training depends on high-speed interconnects like NVLink and network architectures designed to minimize communication latency between GPUs.
#8about 3 minutes
Accessing global GPU capacity with DGX Cloud Lepton
NVIDIA's DGX Cloud Lepton is a marketplace connecting developers to a global network of cloud partners for scalable, on-demand GPU compute.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
00:53 MIN
The rise of general-purpose GPU computing
Accelerating Python on GPUs
20:32 MIN
Accessing software, models, and training resources
Accelerating Python on GPUs
00:48 MIN
The evolution of GPUs from graphics to AI computing
Accelerating Python on GPUs
18:20 MIN
NVIDIA's platform for the end-to-end AI workflow
Trends, Challenges and Best Practices for AI at the Edge
28:04 MIN
Building a cloud architecture for large-scale ML
Geometric deep learning for drug discovery
07:36 MIN
Highlighting impactful contributions and the rise of open models
Open Source: The Engine of Innovation in the Digital Age
01:11 MIN
How GPUs evolved from graphics to AI powerhouses
Accelerating Python on GPUs
24:21 MIN
The future of computing requires scaling out to data centers
Coffee with Developers - Stephen Jones - NVIDIA
Featured Partners
Related Videos
WWC24 - Ankit Patel - Unlocking the Future Breakthrough Application Performance and Capabilities with NVIDIA
Ankit Patel
A Deep Dive on How To Leverage the NVIDIA GB200 for Ultra-Fast Training and Inference on Kubernetes
Kevin Klues
Efficient deployment and inference of GPU-accelerated LLMs
Adolf Hohl
Unveiling the Magic: Scaling Large Language Models to Serve Millions
Patrick Koss
How AI Models Get Smarter
Ankit Patel
AI Factories at Scale
Thomas Schmidt
Exploring LLMs across clouds
Tomislav Tipurić
Generative AI power on the web: making web apps smarter with WebGPU and WebNN
Christian Liebel
From learning to earning
Jobs that call for the skills explored in this talk.

Senior AI Software Developer & Mentor
Dynatrace
Linz, Austria
Senior
Java
TypeScript
AI Frameworks
Agile Methodologies


Senior Software Architect - Deep Learning and HPC Communications
NVIDIA Corporation
Remote
Senior
C++
Linux
Node.js
PyTorch
+1




Cloud & Reliability Engineer - Fokus Machine Learning & Generative AI
Barmer
Bash
Docker
Grafana
Terraform
Prometheus
+4

