(Senior) Data Engineer, Data & Analytics, Group Digital - Madrid

IKEA
Amsterdam, Netherlands
8 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Municipality of San Sebastián de los Reyes, Spain

Tech stack

API
Airflow
Data analysis
Google BigQuery
Continuous Integration
Document-Oriented Databases
Data Flow Control
Identity and Access Management
Python
Network Planning and Design
Operational Databases
SQL Databases
Management of Software Versions
Data Processing
Snowflake
Kafka
Terraform
Network Optimization

Job description

As the lead data engineer embedded in a digital product squad, your mission is to deliver trustworthy, timely, and well-modeled data that powers expansion strategy products used by markets worldwide., * Build Expansion360, the expansion data platform - Architect and operate Expansion360 as Ingka's new, trusted data asset to power prioritized expansion use cases (Location Selection, Store Design, Space Optimization, Investment Management, Project Portfolio Management). Design scalable batch and streaming pipelines on GCP (BigQuery, Dataflow/Beam, Pub/Sub, Composer/dbt) to ingest and harmonize the needed internal and external data to build it. Define canonical models, shared schemas, and data contracts as the single source of truth. Map and govern the needed data domains (store profile & location; sales & cost; market & competition; design & layout; financial & cost; project & process; customer & loyalty; store & market fulfillment capabilities) with SLAs and automated quality checks. Run gap analyses, design acquisition/procurement strategies (third-party data, partnerships, collection pipelines), and close the gaps-backed by robust metadata, lineage, and documentation.

  • Productionize Expansion intelligence - Enable interactive maps and location analytics by enabling geospatial processing at scale (e.g., BigQuery GIS, H3/S2 indexing, isochrones, nearest neighbor joins). Optimize for fast, exploratory queries and heavy analytical workloads.
  • Enable decision and simulation engines - Deliver curated marts and APIs that power scenario planning (e.g., new store P&L, cannibalization/halo, space ROI) and product features (e.g., network optimization simulators, store layout intelligence). Partner with analysts to encode metric definitions and with Data Scientists to serve model inputs/outputs reliably.
  • Engineer for reliability and trust - Implement CI/CD for data (dbt tests, contract tests), observability (data quality SLAs, lineage, anomaly detection), access policies (IAM, row/column security), and cost controls. Own SLAs/SLOs for critical datasets and iterate on performance (partitioning, clustering, materializations).
  • Harden the foundations - Contribute to shared libraries, templates, and infrastructure-as-code (Terraform) to standardize how we build, deploy, and monitor pipelines. Document data contracts and ensure privacy/security best practices are built-in.

Requirements

Do you have experience in Terraform?, We believe the best engineers are systems thinkers and product builders at heart. We would be excited to talk to you if this describes you:

  • Your Technical Foundation - You have 5+ years of hands-on building production data systems. You design and operate batch and streaming pipelines on cloud platforms (GCP preferred) using tools like BigQuery, Dataflow/Beam, Pub/Sub (or Kafka), Cloud Composer/Airflow, and dbt. You're fluent in SQL and write production-grade Python/Scala for data processing and orchestration. You understand data modeling (star/snowflake, vault), partitioning, clustering, and performance at TB-PB scale.
  • How You Solve Problems - You turn ambiguous data needs into robust, observable data products with clear SLAs. You balance speed and rigor, apply 80/20 thinking to prioritize the highest-value work, and design for evolution (schema versioning, data contracts, contract testing). You're comfortable with messy external data (mobility, demographics, POIs) and geospatial datasets, and you bring strong judgment on trade-offs between batch vs. streaming, precompute vs. on-demand, and storage formats.
  • How You Drive Product Success - You build for users, not just for pipelines. You co-define metrics and semantics with analysts and PMs, instrument event streams, and ship data features that enable scenario simulators, interactive maps, and network planning tools. You partner with Data Scientists to productionize features, models, and feature stores with reproducible pipelines, CI/CD, and monitoring.
  • How You Scale Your Impact - You operate with a founder mindset: automate everything that repeats, codify standards, and raise the bar for reliability and quality (testing, lineage, observability). You champion governance and privacy by design (GDPR, PII handling, access controls), mentor engineers, and cultivate strong cross-functional collaboration across Product, Engineering, and Data Science.

This is our wish list! If you don't recognize yourself in every one of these points, you might still be an excellent candidate for the role. We are looking for exceptional individuals and like to think long-term about investing in people's development together with us., Please apply with your application in English; otherwise, we will not be able to process your application.

About the company

We’re a diverse group of down-to-earth, straightforward people with a passion for life at home. We come from all over the world with a vision to inspire and enable people to live more sustainable and healthy lives. Our work is all based on a spirit of togetherness and enthusiasm with care for people and planet. We work hard but we have fun doing it. Together we build a diverse, inclusive, open and honest work environment and are always looking for people who share our positive attitude and values.

Apply for this position