Senior Data Engineer @BENELUX
SquareDev
Brussels, Belgium
3 days ago
Role details
Contract type
Permanent contract Employment type
Full-time (> 32 hours) Working hours
Regular working hours Languages
English Experience level
SeniorJob location
Brussels, Belgium
Tech stack
Artificial Intelligence
Airflow
Azure
Google BigQuery
Databases
Continuous Integration
Data Infrastructure
ETL
DevOps
Distributed Systems
Data Flow Control
Python
SQL Databases
Management of Software Versions
Large Language Models
Snowflake
Spark
GIT
Microsoft Fabric
Data Lake
PySpark
Kubernetes
Data Management
Terraform
Data Pipelines
Docker
Databricks
Job description
If your answer ticks all the boxes, then maybe we can work together.
- You have a curious mind - You won't understand what we're talking about if you don't.
- You want to learn more around technology - You won't survive if you don't.
- You want to make the world a bit better.
We happen to be just like that as well. We like hacking things here and there (you included) and create scalable solutions that bring value to the world., * Building and maintaining data lakes / warehouses (One Lake, BigQuery, Delta Lake).
- Developing and optimizing ETL/ELT workflows using tools like Fabric, Spark Jobs, dbt, Airflow, or Prefect.
- Using / Managing cloud data infrastructure on Azure.
- Ensuring data quality, observability, and governance across all pipelines.
- Working closely with data scientists and software engineers to deploy and maintain AI-ready datasets.
Requirements
Do you have experience in Terraform?, * Designing and implementing data pipelines (batch and streaming) for analytics and AI workloads. The tools used will be python, SQL in Microsoft Fabric as well as low code tools in the Fabric suite., * At least 3 years of relevant work experience.
- Strong experience in SQL and Python (PySpark or similar).
- Hands-on experience with data modeling, ETL frameworks, and data orchestration tools.
- Familiarity with distributed systems and modern data platforms (Spark, Databricks, Fabric, Snowflake, or BigQuery). Fabric will be preferred.
- Understanding of data lifecycle management, versioning, and data testing.
- Solid grasp of Git and CI/CD workflows.
- Strong communication skills in English.
Nice to have:
- Experience with Microsoft Fabric, Data Factory, Dataflow Gen2
- Knowledge of vector databases (pgvector, Pinecone, Milvus) or semantic search pipelines.
- Interest / knowledge in LLMs, AI pipelines.
- Familiarity with data catalogs, lineage tools, or dbt tests.
- DevOps familiarity (Docker, Kubernetes, Terraform). Certifications in Azure, Fabric or similar platforms.