How E.On productionizes its AI model & Implementation of Secure Generative AI.
Stop just prompt engineering. E.On reveals how they build production-grade LLM guards to secure their AI and solve core business challenges.
#1about 2 minutes
Defining the roles of data scientists and MLOps engineers
MLOps engineers build secure, automated infrastructure to help data scientists move models from proof-of-concept to production in weeks instead of months.
#2about 2 minutes
Implementing the data as code concept for ML
E.On uses a "data as code" approach with versioned Python libraries to provide data scientists with abstracted, quality-checked data frames.
#3about 3 minutes
Using LLMs to discover datasets and manage metadata
LLMs can query enterprise metadata, helping users find relevant data sources and even generate SQL queries through natural language conversations.
#4about 1 minute
Powering website search with generative AI
E.On is replacing traditional keyword search on its website with a GPT-powered system that provides crisp, human-like answers based on a broad knowledge base.
#5about 2 minutes
Using LLMs to understand and navigate codebases
Developers can use a combination of LangChain, vector databases, and OpenAI to ask natural language questions about a large codebase for faster debugging.
#6about 1 minute
Architecture for analyzing contact center call recordings
An architecture using speech-to-text and Azure OpenAI analyzes customer call recordings to extract sentiment, verify GDPR consent, and uncover business insights.
#7about 2 minutes
Automating email processing with intent classification
To manage millions of customer emails, an automated system uses OpenAI for intent classification to suggest accurate and fast responses for call center agents.
#8about 3 minutes
Securely connecting generative AI to enterprise data
An architecture combining a knowledge base, cognitive search, and embeddings allows GPT to securely answer questions using private enterprise data, protected by LLM guards.
#9about 1 minute
Overview of generative AI applications at E.On
A summary of key application areas for foundation models, including real-time carbon footprinting, hyper-personalization, and smart data analysis.
Related jobs
Jobs that call for the skills explored in this talk.
MLops – Deploying, Maintaining And Evolving Machine Learning Models in ProductionWelcome to this issue of the WeAreDevelopers Live Talk series. This article recaps an interesting talk by Bas Geerdink who gave advice on MLOps.About the speaker:Bas is a programmer, scientist, and IT manager. At ING, he is responsible for the Fast...
Benedikt Bischof
MLOps And AI Driven DevelopmentWelcome to this issue of the WeAreDevelopers Dev Talk Recap series. This article recaps an interesting talk by Natalie Pistunovic who spoke about the development of AI and MLOps. What you will learn:How the concept of AI became an academic field and ...
Chris Heilmann
With AIs wide open - WeAreDevelopers at All Things Open 2025Last week our VP of Developer Relations, Chris Heilmann, flew to Raleigh, North Carolina to present at All Things Open . An excellent event he had spoken at a few times in the past and this being the “Lucky 13” edition, he didn’t hesitate to come and...
Eli McGarvie
13 AI Tools You Have to TryFirst, it was NFTs, then it was Web3, and now it’s generative AI… it’s probably time to stop collecting pictures of monkeys and kitties. Chatbots and generative AI are the next big thing. This time we’ve jumped on a trend that has real-world applicat...
From learning to earning
Jobs that call for the skills explored in this talk.