GCP Data Architect
Role details
Job location
Tech stack
Job description
We are seeking an experienced Cloud Data Loading Architect to design, build, and optimise automated pipelines that ingest structured, semistructured, and unstructured datasets into Google Cloud Platform (GCP), specifically BigQuery. This role will lead endtoend data ingestion design-from source discovery and schema mapping, through transformation and data quality, to scalable, secure loads into cloudnative analytical warehouses. The ideal candidate combines strong cloud engineering skills with handson data integration experience and a deep understanding of BigQuery performance optimisation., · Design and implement highthroughput, faulttolerant ingestion pipelines for batch and streaming data landing in BigQuery. · Lead data ingestion architecture patterns using Cloud Storage (GCS), Dataflow, Dataproc, Composer (Airflow), Pub/Sub, BigQuery Storage Write API, and related services.
· Define data loading frameworks, mapping rules, schema evolution strategy, and metadata management. · Create reusable ingestion blueprints that ensure governance, lineage, and auditability. · Establish data quality checks, validation rules, reconciliation logic, and SLAs. · Optimise BigQuery cost, storage, partitioning, clustering, and access patterns. · Collaborate with security & platform teams to ensure IAM, service accounts, VPCSC, and encryption policies are fully applied. · Automate CI/CD deployments for ingestion pipelines using Cloud Build, GitHub, GitLab or Jenkins. · Produce detailed technical documentation and coach engineering squads in cloud ingestion standards. · Troubleshoot ingestion failures, performance bottlenecks, and crossplatform data integration issues.
Requirements
-
Deep Expertise in Google Cloud Data Services BigQuery, GCS, Dataflow (Apache Beam), Pub/Sub, Dataproc, Cloud Composer, Storage Write API.
-
Data Ingestion Engineering Mastery Handson experience designing frameworks to load data from APIs, files, databases, event streams, and mainframe/legacy systems into cloud stores.
-
Strong SQL & BigQuery Optimisation Skills Partitioning, clustering, materialised views, costefficient query design, columnar storage understanding.
-
ETL/ELT Architecture Knowledge Experience building transformation pipelines using Airflow, Dataflow, dbt, or equivalent orchestration tools.
-
File & Format Proficiency
Ability to work with Parquet, Avro, ORC, JSON, CSV, nested/repeated structures, and schema evolution.
-
Strong Python and/or Java Skills Used to build Dataflow pipelines, ingestion utilities, automation scripts.
-
Cloud Security & Governance Awareness IAM roles, leastprivilege models, VPCSC, service accounts, artifact signing, audit logging.
-
DevOps & CI/CD Familiarity Cloud Build, GitHub Actions, Terraform, Cloud Deployment Manager or Pulumi.
-
Data Quality & Observability Mindset Experience implementing validation frameworks, anomaly detection, reconciliation rules, logging/monitoring (e.g., Cloud Logging, Cloud Monitoring).
-
Excellent Architectural Communication Skills Ability to document, diagram, and communicate ingestion patterns to stakeholders at technical and nontechnical levels.
Benefits & conditions
Pay: £500.00-£600.00 per day