AI Engineer (NL25)
Role details
Job location
Tech stack
Job description
We are looking for a highly skilled AI Engineer to join our team. In this role, you will be responsible for designing, building, and deploying scalable AI and Machine Learning solutions. You will work at the intersection of Data Engineering, DevOps, and Machine Learning to create robust Generative AI solutions and traditional ML pipelines. You will leverage cutting-edge frameworks and cloud services to solve complex business problems, ensuring our AI systems are reliable, efficient, and secure. Key responsibilities
- Design and implement end-to-end AI/ML pipelines, from data ingestion to model deployment and monitoring
- Build and optimise Generative AI applications using RAG (Retrieval-Augmented Generation) architectures and LLM frameworks
- Develop and maintain scalable data processing workflows using modern big data tools
- Implement MLOps best practices for model lifecycle management, versioning and automated deployment
- Collaborate with data scientists and stakeholders to translate business requirements into technical solutions
- Ensure data quality, governance, and security across all AI and data platforms, * Mosaic AI: Utilisation of Mosaic AI tools for Large Language Models, Vector Search integration, and Model Serving endpoints
- MLflow: Advanced usage of MLflow within Databricks for full lifecycle management of ML models and experiments
- Unity Catalog: Implementation of Unity Catalog for centralised governance of data and AI assets (models, functions)
- Jobs & DLT: Building reliable production pipelines using Databricks Jobs, Workflows, and Delta Live Tables (DLT)
- Optimisation: Performance tuning of Spark SQL and PySpark jobs to ensure cost-effectiveness and low latency
- Automation: Scripting and automation using the Databricks CLI and REST APIs
Certifications (nice to have)Core & General
- NVIDIA Certified Associate: AI in the Data Center or similar GenAI certifications
- DeepLearning.AI: Start or Advance Your Career in AI: Specialised courses (e.g., Generative AI with LLMs)
- AWS Certified Machine Learning - Specialty: Validates expertise in building, training, tuning and deploying ML models on AWS
- AWS Certified Solutions Architect - Associate
- Microsoft Certified: Azure AI Engineer Associate (AI-102): Validates ability to build, manage, and deploy AI solutions leveraging Azure Cognitive Services and Azure Applied AI Services
- Microsoft Certified: Azure Data Scientist Associate (DP-100): Focuses on designing and implementing a data science solution on Azure
Data Platform: Databricks
- Databricks Certified Machine Learning Professional: Demonstrates ability to use Databricks Machine Learning and MLflow for production ML tasks
- Databricks Certified Generative AI Engineer Associate: Validates understanding of Generative AI concepts and the ability to develop LLM applications on Databricks
- Databricks Certified Data Engineer Professional: For candidates with a stronger data engineering background
Requirements
Bachelor's degree in Computer Science or related field, * A minimum of 5 years professional experience, with at least 2 years' experience within the AI/ML space, and a year in Generative AI highly desirable
Skills and ExperienceCore skills, tools and frameworks
-
Programming mastery: Expert-level proficiency in Python, with a strong grasp of Object-Oriented Programming (OOP) principles, design patterns, and asynchronous programming
-
AI & LLM frameworks: Experience with modern AI frameworks such as LangChain, Langflow, and AutoGen. Ability to build agents and orchestrate complex LLM workflows
-
GenAI architecture: Deep understanding of Generative AI patterns, including RAG (Retrieval-Augmented Generation), Vector Search implementation (e.g., Pinecone, Chroma, Milvus), and advanced Prompt Engineering techniques
-
ML engineering & MLOps: Practical experience with MLOps tools like MLflow for experiment tracking and model registry, implementation of Feature Stores, and Model Serving infrastructures (e.g., TFServing, TorchServe, KServe)
-
Big Data processing: Proficiency in PySpark, Spark SQL, and Delta Lake for handling large-scale datasets and optimising data pipelines
-
Version control & CI/CD: Background in Git workflows and CI/CD pipelines using tools like GitHub Actions or Azure DevOps for automated testing and deployment
-
Orchestration: Experience scheduling and managing complex workflows using Apache Airflow, Databricks Workflows, or Delta Live Tables (DLT)
-
Data architecture: Familiarity with modern data architecture patterns such as the Medallion Architecture (Bronze/Silver/Gold) and data modelling techniques (ETL/ELT)
-
Data governance: Knowledge of data governance frameworks, specifically Unity Catalog, including data lineage, access controls, and security policies
-
Streaming: Experience with real-time data processing technologies like Apache Kafka, AWS Kinesis, or Spark Structured Streaming
-
Data quality: Implementation of data quality checks and validation using libraries like Great Expectations or Deequ
-
BI & visualisation: Ability to create dashboards and visualise insights using tools like PowerBI, Tableau, or Databricks SQL Dashboards
-
Storage: Comprehensive knowledge of Amazon S3 for data lake storage, lifecycle policies, and security configurations
-
IAM & security: Proficiency in AWS IAM (Identity and Access Management) for managing roles, policies, and secure access to resources
-
Networking: Understanding of VPCs, subnets, security groups, and general AWS networking concepts to ensure secure deployment
-
AI/ML services: Hands-on experience with managed AI services such as Amazon Bedrock for building GenAI applications and Amazon SageMaker for model training and deployment
-
Streaming: Experience configuring and managing Amazon Kinesis for real-time data ingestion and processing
-
Storage: Expertise in Azure Data Lake Storage Gen2 (ADLS Gen2), including hierarchical namespaces and ACLs
-
Azure OpenAI Service: Experience deploying and managing LLMs via Azure OpenAI Service, including model fine-tuning and content filtering
-
AI Search: Implementation of Azure AI Search (formerly Cognitive Search) with vector capabilities for semantic search applications
-
Core Azure knowledge: Solid understanding of core Azure infrastructure, resource management (ARM templates / Bicep), and monitoring
-
Networking & security: Proficiency in Azure networking (VNETs, Private Endpoints) and security best practices (Managed Identities, Key Vault), We are open to training internal candidates who demonstrate strong engineering fundamentals and a passion for data. Ideal internal candidates might currently be in the following roles: Python Back-end Engineers
-
Background: Strong experience in Python, API development (FastAPI/Flask), and system architecture
-
Gap to bridge: Need to learn specific AI frameworks (LangChain), ML concepts (embeddings, vectors), and data-centric tools (Spark, Vector DBs)
-
Background: Proficient in Spark, SQL, pipelines, and cloud infrastructure
-
Gap to bridge: Need to deepen knowledge in Model Serving, MLflow, LLM architectures (RAG), and non-deterministic software behavior (prompt engineering)
Data Scientists
- Background: Strong in statistics, model building, and experimentation in notebooks
- Gap to bridge: Need to improve software engineering practices (CI/CD, testing, modular code) and learn production deployment patterns (serving, monitoring)