Senior Data Engineer @BENELUX

SquareDev
Brussels, Belgium
3 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Brussels, Belgium

Tech stack

Artificial Intelligence
Airflow
Azure
Google BigQuery
Databases
Continuous Integration
Data Infrastructure
ETL
DevOps
Distributed Systems
Data Flow Control
Python
SQL Databases
Management of Software Versions
Large Language Models
Snowflake
Spark
GIT
Microsoft Fabric
Data Lake
PySpark
Kubernetes
Data Management
Terraform
Data Pipelines
Docker
Databricks

Job description

If your answer ticks all the boxes, then maybe we can work together.

  • You have a curious mind - You won't understand what we're talking about if you don't.
  • You want to learn more around technology - You won't survive if you don't.
  • You want to make the world a bit better.

We happen to be just like that as well. We like hacking things here and there (you included) and create scalable solutions that bring value to the world., * Building and maintaining data lakes / warehouses (One Lake, BigQuery, Delta Lake).

  • Developing and optimizing ETL/ELT workflows using tools like Fabric, Spark Jobs, dbt, Airflow, or Prefect.
  • Using / Managing cloud data infrastructure on Azure.
  • Ensuring data quality, observability, and governance across all pipelines.
  • Working closely with data scientists and software engineers to deploy and maintain AI-ready datasets.

Requirements

Do you have experience in Terraform?, * Designing and implementing data pipelines (batch and streaming) for analytics and AI workloads. The tools used will be python, SQL in Microsoft Fabric as well as low code tools in the Fabric suite., * At least 3 years of relevant work experience.

  • Strong experience in SQL and Python (PySpark or similar).
  • Hands-on experience with data modeling, ETL frameworks, and data orchestration tools.
  • Familiarity with distributed systems and modern data platforms (Spark, Databricks, Fabric, Snowflake, or BigQuery). Fabric will be preferred.
  • Understanding of data lifecycle management, versioning, and data testing.
  • Solid grasp of Git and CI/CD workflows.
  • Strong communication skills in English.

Nice to have:

  • Experience with Microsoft Fabric, Data Factory, Dataflow Gen2
  • Knowledge of vector databases (pgvector, Pinecone, Milvus) or semantic search pipelines.
  • Interest / knowledge in LLMs, AI pipelines.
  • Familiarity with data catalogs, lineage tools, or dbt tests.
  • DevOps familiarity (Docker, Kubernetes, Terraform). Certifications in Azure, Fabric or similar platforms.

Apply for this position