(Senior) Data Engineer, Data & Analytics, Group Digital - Madrid

IKEA
Amsterdam, Netherlands
2 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Municipality of Madrid, Spain

Tech stack

API
Airflow
Data analysis
Google BigQuery
Continuous Integration
Document-Oriented Databases
Data Flow Control
Identity and Access Management
Python
Network Planning and Design
Operational Databases
SQL Databases
Management of Software Versions
Data Processing
Snowflake
Kafka
Terraform
Network Optimization

Job description

  • How You Drive Product Success-You build for users, not just pipelines. You co-define metrics and semantics with analysts and PMs, instrument event streams, and ship data features that enable scenario simulators, interactive maps, and network planning tools. You partner with Data Scientists to productionise features, models, and feature stores with reproducible pipelines, CI/CD, and monitoring.
  • How You Scale Your Impact-You operate with a founder mindset: automate everything that repeats, codify standards, and raise the bar for reliability and quality (testing, lineage, observability). You champion governance and privacy by design (GDPR, PII handling, access controls), mentor engineers, and cultivate strong cross-functional collaboration across Product, Engineering, and Data Science.

This is our wish list! If you don't recognise yourself in every one of these points, you might still be an excellent candidate for the role. We are looking for exceptional individuals and like to think long-term about investing in people's development together with us. What You'll Be Doing Day to Day

  • Build Expansion360-Architect and operate Expansion360 as Ingka's new, trusted data asset to power prioritised expansion use cases (Location Selection, Store Design, Space Optimisation, Investment Management, Project Portfolio Management). Design scalable batch and streaming pipelines on GCP (BigQuery, Dataflow/Beam, Pub/Sub, Composer/dbt) to ingest and harmonise the needed internal and external data. Define canonical models, shared schemas, and data contracts as the single source of truth. Map and govern the needed data domains with SLAs and automated quality checks. Run gap analyses, design acquisition/procurement strategies (third-party data, partnerships, collection pipelines), and close the gaps-backed by robust metadata, lineage, and documentation.
  • Productionise Expansion Intelligence-Enable interactive maps and location analytics by enabling geospatial processing at scale (e.g., BigQuery GIS, H3/S2 indexing, isochrones, nearest-neighbor joins). Optimise for fast, exploratory queries and heavy analytical workloads.
  • Enable Decision and Simulation Engines-Deliver curated marts and APIs that power scenario planning (e.g., new store P&L, cannibalisation/halo, space ROI) and product features (e.g., network optimisation simulators, store layout intelligence). Partner with analysts to encode metric definitions and with Data Scientists to serve model inputs/outputs reliably.
  • Engineer for Reliability and Trust-Implement CI/CD for data (dbt tests, contract tests), observability (data quality SLAs, lineage, anomaly detection), access policies (IAM, row/column security), and cost controls. Own SLAs/SLOs for critical datasets and iterate on performance (partitioning, clustering, materialisations).
  • Harden the Foundations-Contribute to shared libraries, templates, and infrastructure-as-code (Terraform) to standardise how we build, deploy, and monitor pipelines. Document data contracts and ensure privacy/security best practices are built-in.

Requirements

  • Technical Foundation-You have 5+ years of hands-on building production data systems. You design and operate batch and streaming pipelines on cloud platforms (GCP preferred) using tools like BigQuery, Dataflow/Beam, Pub/Sub (or Kafka), Cloud Composer/Airflow, and dbt. You're fluent in SQL and write production-grade Python/Scala for data processing and orchestration. You understand data modeling (star/snowflake, vault), partitioning, clustering, and performance at TB-PB scale.
  • How You Solve Problems-You turn ambiguous data needs into robust, observable data products with clear SLAs. You balance speed and rigor, apply 80/20 thinking to prioritise the highest-value work, and design for evolution (schema versioning, data contracts, contract testing). You're comfortable with messy external data (mobility, demographics, POIs) and geospatial datasets, and you bring strong judgement on trade-offs between batch vs. streaming, pre-compute vs. on-demand, and storage formats., Please apply with your application in English; otherwise, we will not be able to process your application.

About the company

We’re a diverse group of down-to-earth, straightforward people with a passion for life at home. We come from all over the world with a vision to inspire and enable people to live more sustainable and healthy lives. Our work is all based on a spirit of togetherness and enthusiasm with care for people and planet. We work hard but we have fun doing it. Together we build a diverse, inclusive, open and honest work environment and are always looking for people who share our positive attitude and values.

Apply for this position