Ekaterina Sirazitdinova

Trends, Challenges and Best Practices for AI at the Edge

How do you run powerful AI on resource-constrained devices? Learn the key optimization techniques, from synthetic data and quantization to high-performance runtimes.

Trends, Challenges and Best Practices for AI at the Edge
#1about 4 minutes

Defining AI at the edge and its industry applications

AI at the edge involves running computations on devices near the data source, transforming industries like manufacturing, retail, and healthcare.

#2about 1 minute

Understanding the unique constraints of edge devices

Edge devices differ from data centers due to their limited compute power, smaller storage capacity, and restricted power consumption.

#3about 2 minutes

Overcoming the primary challenges of edge AI development

Developers must solve for three main challenges: achieving high model accuracy, ensuring real-time throughput, and managing deployment at scale.

#4about 1 minute

Using synthetic data to improve model accuracy

Synthetic data helps improve model accuracy by providing diverse training examples, covering rare corner cases, and reducing expensive manual labeling.

#5about 3 minutes

Optimizing models with quantization and network pruning

Model performance can be significantly improved by using quantization to reduce numerical precision and network pruning to remove unnecessary neurons.

#6about 4 minutes

Advanced techniques for boosting inference performance

Further performance gains can be achieved through network graph optimizations, kernel auto-tuning, dynamic tensor memory, and multistream concurrent execution.

#7about 1 minute

NVIDIA's platform for the end-to-end AI workflow

NVIDIA provides a comprehensive software platform to support the entire AI productization cycle, from data collection and training to optimization and deployment.

#8about 2 minutes

Using Replicator and pre-trained models for development

NVIDIA Replicator generates synthetic data for training, while the NGC catalog offers a wide range of pre-trained models to accelerate development.

#9about 2 minutes

Training and fine-tuning models with the TAO Toolkit

The NVIDIA TAO Toolkit is a zero-coding framework that simplifies training, fine-tuning, pruning, and quantization of AI models.

#10about 2 minutes

Deploying models with TensorRT and Triton Inference Server

NVIDIA TensorRT optimizes models for high-performance inference, while Triton Inference Server provides a flexible solution for serving models at scale.

#11about 2 minutes

Building video analytics pipelines with DeepStream SDK

The NVIDIA DeepStream SDK, built on GStreamer, enables the creation of efficient, GPU-accelerated video analytics pipelines with zero memory copies.

#12about 2 minutes

Matching edge AI challenges with NVIDIA's solutions

A summary of how NVIDIA's tools like Replicator, TAO Toolkit, TensorRT, and DeepStream address the core challenges of accuracy, performance, and deployment.

Related jobs
Jobs that call for the skills explored in this talk.

Featured Partners

From learning to earning

Jobs that call for the skills explored in this talk.