IT Engineer IV

Randstad
Amsterdam, Netherlands
2 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Junior

Job location

Amsterdam, Netherlands

Tech stack

Amazon Web Services (AWS)
Data analysis
Azure
Continuous Integration
Data Governance
Data Infrastructure
ETL
Dimensional Modeling
Distributed Data Store
Amazon DynamoDB
Github
Python
Online Analytical Processing
Online Transaction Processing
Systems Development Life Cycle
SQL Databases
Cloud Platform System
Cloudformation
Data Lake
PySpark
Data Lineage
Star Schema
Software Version Control
Data Pipelines
Databricks
Go

Job description

Description of activities:

We are looking for a data engineer at BUUT to design, build, and maintain a scalable, general-purpose data lake and data pipelines that enable the L&I (learning&insights) team to generate actionable insights. This role will ensure data accessibility, quality, and performance for both analytical and reporting needs, reducing dependency on multiple engineers and creating a unified, efficient data infrastructure.

  • With the following results:
  1. Data Pipeline Development
  • Design and implement at least 2-3 robust data pipelines that support L&I analytics and reporting needs.
  • Ensure pipelines are automated, tested, and monitored for reliability.
  1. Enable Analytics & Insights
  • Provide usable datasets and query capabilities for L&I to generate insights.
  • Deliver at least one analytics feature or dashboard powered by the new data infrastructure.
  1. Operational Excellence
  • Implement CI/CD workflows for data jobs (GitHub Actions).
  • Set up basic monitoring and alerting for pipeline health and data quality.
  1. Data Governance & Quality
  • Define and apply data quality checks (row-level and aggregate).
  • Establish data lineage documentation for key pipelines.
  • Relevant knowledge skills & competences:

Must-haves:

Tools & Platforms:

  • Version Control & CI/CD: GitHub, GitHub Actions
  • At least one major cloud platform: AWS, Azure, or Databricks

Languages: Python, SQL

Experience: Minimum 3 years creating and maintaining data pipelines

Core Knowledge:

  • Data pipelines (ETL), orchestration (batch vs streaming)
  • Data modeling (star schema, dimensional modeling, SCD)
  • OLTP vs OLAP concepts
  • Testing (unit, integration, E2E)
  • Data governance basics (lineage, quality checks)

Nice-to-haves:

Tools & Platforms:

  • AWS services: Glue, Athena, DynamoDB, Step Functions
  • Languages: PySpark, Golang

Patterns & Techniques:

  • Infrastructure as Code (AWS CloudFormation)
  • Table formats: Iceberg / Delta / Hudi
  • Schema evolution, reprocessing, monitoring
  • Medallion architecture
  • Distributed data processing

Experience:

  • Improving SDLC for data teams (validation, testing automation)
  • Generating insights for end-users (e.g., personalization)

,

Description of activities:

We are looking for a data engineer at BUUT to design, build, and maintain a scalable, general-purpose data lake and data pipelines that enable the L&I (learning&insights) team to generate actionable insights. This role will ensure data accessibility, quality, and performance for both analytical and reporting needs, reducing dependency on multiple engineers and creating a unified, efficient data infrastructure.

  • With the following results:
  1. Data Pipeline Development
  • Design and implement at least 2-3 robust data pipelines that support L&I analytics and reporting needs.
  • Ensure pipelines are automated, tested, and monitored for reliability.
  1. Enable Analytics & Insights
  • Provide usable datasets and query capabilities for L&I to generate insights.
  • Deliver at least one analytics feature or dashboard powered by the new data infrastructure.
  1. Operational Excellence
  • Implement CI/CD workflows for data jobs (GitHub Actions).
  • Set up basic monitoring and alerting for pipeline health and data quality.
  1. Data Governance & Quality
  • Define and apply data quality checks (row-level and aggregate).
  • Establish data lineage documentation for key pipelines.
  • Relevant knowledge skills & competences:

Must-haves:

Tools & Platforms:

  • Version Control & CI/CD: GitHub, GitHub Actions
  • At least one major cloud platform: AWS, Azure, or Databricks

Languages: Python, SQL

Experience: Minimum 3 years creating and maintaining data pipelines

Core Knowledge:

  • Data pipelines (ETL), orchestration (batch vs streaming)
  • Data modeling (star schema, dimensional modeling, SCD)
  • OLTP vs OLAP concepts
  • Testing (unit, integration, E2E)
  • Data governance basics (lineage, quality checks)

Nice-to-haves:

Tools & Platforms:

  • AWS services: Glue, Athena, DynamoDB, Step Functions
  • Languages: PySpark, Golang

Patterns & Techniques:

  • Infrastructure as Code (AWS CloudFormation)
  • Table formats: Iceberg / Delta / Hudi
  • Schema evolution, reprocessing, monitoring
  • Medallion architecture
  • Distributed data processing

Experience:

  • Improving SDLC for data teams (validation, testing automation)
  • Generating insights for end-users (e.g., personalization)

Requirements

  • Version Control & CI/CD: GitHub, GitHub Actions
  • At least one major cloud platform: AWS, Azure, or Databricks

Languages: Python, SQL

Experience: Minimum 3 years creating and maintaining data pipelines

Core Knowledge:

  • Data pipelines (ETL), orchestration (batch vs streaming)
  • Data modeling (star schema, dimensional modeling, SCD)
  • OLTP vs OLAP concepts
  • Testing (unit, integration, E2E)
  • Data governance basics (lineage, quality checks)

Nice-to-haves:

Tools & Platforms:

  • AWS services: Glue, Athena, DynamoDB, Step Functions
  • Languages: PySpark, Golang

Patterns & Techniques:

  • Infrastructure as Code (AWS CloudFormation)
  • Table formats: Iceberg / Delta / Hudi
  • Schema evolution, reprocessing, monitoring
  • Medallion architecture
  • Distributed data processing

Experience:

  • Improving SDLC for data teams (validation, testing automation)
  • Generating insights for end-users (e.g., personalization)

Apply for this position