(Senior) Data Engineer, Data & Analytics, Group Digital - Madrid
Role details
Job location
Tech stack
Job description
As the lead data engineer embedded in a digital product squad, your mission is to deliver trustworthy, timely, and well-modeled data that powers expansion strategy products used by markets worldwide., * Build Expansion360, the expansion data platform - Architect and operate Expansion360 as Ingka's new, trusted data asset to power prioritized expansion use cases (Location Selection, Store Design, Space Optimization, Investment Management, Project Portfolio Management). Design scalable batch and streaming pipelines on GCP (BigQuery, Dataflow/Beam, Pub/Sub, Composer/dbt) to ingest and harmonize the needed internal and external data to build it. Define canonical models, shared schemas, and data contracts as the single source of truth. Map and govern the needed data domains (store profile & location; sales & cost; market & competition; design & layout; financial & cost; project & process; customer & loyalty; store & market fulfillment capabilities) with SLAs and automated quality checks. Run gap analyses, design acquisition/procurement strategies (third-party data, partnerships, collection pipelines), and close the gaps-backed by robust metadata, lineage, and documentation.
- Productionize Expansion intelligence - Enable interactive maps and location analytics by enabling geospatial processing at scale (e.g., BigQuery GIS, H3/S2 indexing, isochrones, nearest neighbor joins). Optimize for fast, exploratory queries and heavy analytical workloads.
- Enable decision and simulation engines - Deliver curated marts and APIs that power scenario planning (e.g., new store P&L, cannibalization/halo, space ROI) and product features (e.g., network optimization simulators, store layout intelligence). Partner with analysts to encode metric definitions and with Data Scientists to serve model inputs/outputs reliably.
- Engineer for reliability and trust - Implement CI/CD for data (dbt tests, contract tests), observability (data quality SLAs, lineage, anomaly detection), access policies (IAM, row/column security), and cost controls. Own SLAs/SLOs for critical datasets and iterate on performance (partitioning, clustering, materializations).
- Harden the foundations - Contribute to shared libraries, templates, and infrastructure-as-code (Terraform) to standardize how we build, deploy, and monitor pipelines. Document data contracts and ensure privacy/security best practices are built-in.
Requirements
Do you have experience in Terraform?, We believe the best engineers are systems thinkers and product builders at heart. We would be excited to talk to you if this describes you:
- Your Technical Foundation - You have 5+ years of hands-on building production data systems. You design and operate batch and streaming pipelines on cloud platforms (GCP preferred) using tools like BigQuery, Dataflow/Beam, Pub/Sub (or Kafka), Cloud Composer/Airflow, and dbt. You're fluent in SQL and write production-grade Python/Scala for data processing and orchestration. You understand data modeling (star/snowflake, vault), partitioning, clustering, and performance at TB-PB scale.
- How You Solve Problems - You turn ambiguous data needs into robust, observable data products with clear SLAs. You balance speed and rigor, apply 80/20 thinking to prioritize the highest-value work, and design for evolution (schema versioning, data contracts, contract testing). You're comfortable with messy external data (mobility, demographics, POIs) and geospatial datasets, and you bring strong judgment on trade-offs between batch vs. streaming, precompute vs. on-demand, and storage formats.
- How You Drive Product Success - You build for users, not just for pipelines. You co-define metrics and semantics with analysts and PMs, instrument event streams, and ship data features that enable scenario simulators, interactive maps, and network planning tools. You partner with Data Scientists to productionize features, models, and feature stores with reproducible pipelines, CI/CD, and monitoring.
- How You Scale Your Impact - You operate with a founder mindset: automate everything that repeats, codify standards, and raise the bar for reliability and quality (testing, lineage, observability). You champion governance and privacy by design (GDPR, PII handling, access controls), mentor engineers, and cultivate strong cross-functional collaboration across Product, Engineering, and Data Science.
This is our wish list! If you don't recognize yourself in every one of these points, you might still be an excellent candidate for the role. We are looking for exceptional individuals and like to think long-term about investing in people's development together with us., Please apply with your application in English; otherwise, we will not be able to process your application.
About the company
We’re a diverse group of down-to-earth, straightforward people with a passion for life at home. We come from all over the world with a vision to inspire and enable people to live more sustainable and healthy lives. Our work is all based on a spirit of togetherness and enthusiasm with care for people and planet. We work hard but we have fun doing it. Together we build a diverse, inclusive, open and honest work environment and are always looking for people who share our positive attitude and values.