Anshul Jindal
LLMOps-driven fine-tuning, evaluation, and inference with NVIDIA NIM & NeMo Microservices
#1about 6 minutes
Understanding the GenAI lifecycle and its operational challenges
The continuous cycle of data processing, model customization, and deployment for GenAI applications creates production complexities like a lack of standardized CI/CD and versioning.
#2about 2 minutes
Breaking down the structured stages of an LLMOps pipeline
An effective LLMOps process moves a model from an experimental proof-of-concept through evaluation, pre-production testing, and finally to a production environment.
#3about 4 minutes
Introducing the NVIDIA NeMo microservices and ecosystem tools
NVIDIA provides a suite of tools including NeMo Curator, Customizer, Evaluator, and NIM, which integrate with ecosystem components like Argo Workflows and Argo CD for a complete LLMOps solution.
#4about 4 minutes
Using NeMo Customizer and Evaluator for model adaptation
NeMo Customizer and Evaluator simplify model adaptation through API requests that trigger fine-tuning on custom datasets and benchmark the resulting model's performance.
#5about 3 minutes
Deploying and scaling models with NVIDIA NIM on Kubernetes
NVIDIA NIM packages models into optimized inference containers that can be deployed and auto-scaled on Kubernetes using the NIM operator, with support for multiple fine-tuned adapters.
#6about 4 minutes
Automating complex LLM workflows with Argo Workflows
Argo Workflows enables the creation of automated, multi-step pipelines by stitching together containerized tasks for data processing, model customization, evaluation, and deployment.
#7about 3 minutes
Implementing a GitOps approach for end-to-end LLMOps
Using Git as the single source of truth, Argo CD automates the deployment and management of all LLMOps components, including microservices and workflows, onto Kubernetes clusters.
#8about 3 minutes
Demonstrating the automated LLMOps pipeline in action
A practical demonstration shows how Argo CD manages deployed services and how a data scientist can launch a complete fine-tuning workflow through the Argo Workflows UI, with results tracked in MLflow.
Related jobs
Jobs that call for the skills explored in this talk.
Matching moments
06:19 MIN
Defining LLMOps and understanding its core benefits
From Traction to Production: Maturing your LLMOps step by step
39:32 MIN
Implementing a CI/CD pipeline for your NLP model
Multilingual NLP pipeline up and running from scratch
01:01 MIN
Understanding the role and challenges of MLOps
The Road to MLOps: How Verivox Transitioned to AWS
40:05 MIN
How to assess and advance your LLMOps maturity
From Traction to Production: Maturing your LLMOps step by step
09:27 MIN
Using MLOps infrastructure to implement model governance
Model Governance and Explainable AI as tools for legal compliance and risk management
22:41 MIN
Introducing the Azure AI platform for end-to-end LLMOps
From Traction to Production: Maturing your LLMOps step by step
06:34 MIN
Understanding the machine learning workflow and MLOps
Machine Learning in ML.NET
09:21 MIN
Differentiating between LLMOps and traditional MLOps
From Traction to Production: Maturing your LLMOps step by step
Featured Partners
Related Videos
DevOps for AI: running LLMs in production with Kubernetes and KubeFlow
Aarno Aukia
Adding knowledge to open-source LLMs
Sergio Perez & Harshita Seth
Efficient deployment and inference of GPU-accelerated LLMs
Adolf Hohl
From Traction to Production: Maturing your LLMOps step by step
Maxim Salnikov
MLOps on Kubernetes: Exploring Argo Workflows
Hauke Brammer
MLOps - What’s the deal behind it?
Nico Axtmann
Effective Machine Learning - Managing Complexity with MLOps
Simon Stiebellehner
Self-Hosted LLMs: From Zero to Inference
Roberto Carratalá & Cedric Clyburn
From learning to earning
Jobs that call for the skills explored in this talk.


AI Software Engineer - Big Data Pipelines & ML Automation | Python, C#, C++ Expert | Machine Learning Engineer in Manufacturing
Imnoo
Remote
Senior
C++
ETL
.NET
REST
+26

Data Scientist- Python/MLflow-NLP/MLOps/Generative AI
ITech Consult AG
Azure
Python
PyTorch
TensorFlow
Machine Learning

Machine Learning (MLOps) Engineer
Da Vinci Engineering GmbH
Intermediate
Azure
DevOps
Python
Docker
PyTorch
+6

AI Enablement Engineer - LLM Integration & Technical Empowerment
KUEHNE + NAGEL
Intermediate
API
Python
Docker
Kubernetes
Continuous Integration
+1



AI Engineer / MLOps Engineer
VESTIGAS GmbH
Senior
Azure
Python
Machine Learning
Natural Language Processing

MLOps Engineer für den Bereich Künstliche Intelligenz (Artificial Intelligence)
ROHDE & SCHWARZ GmbH & Co. KG
DevOps
Python
Grafana
Prometheus
Kubernetes
+2