Google Cloud Lead Data Engineer (GCP)
Role details
Job location
Tech stack
Job description
We are seeking a skilled Data Engineer with hands-on experience in Databricks and Google Cloud Platform (GCP) to design, build, and optimize data pipelines and analytics solutions. The ideal candidate will have a strong background in distributed data processing, cloud architecture, and data modeling. This role partners closely with data analysts, data scientists, and business stakeholders to deliver scalable, reliable, and high-quality data products.
Your role
Design, build, and maintain ETL/ELT pipelines using Databricks (PySpark, Delta Lake). Optimize pipelines for performance, cost efficiency, and scalability within GCP. Develop batch and streaming data processes using Spark Streaming, and related technologies. Implement data solutions leveraging GCP services such as BigQuery, Cloud Storage, Dataflow, Cloud Composer, and Vertex AI integrations. Apply best practices for cloud security, IAM configuration, monitoring, and cost management. Build and maintain data models, including dimensional modeling and data vault structures. Implement data quality frameworks, validation rules, and automated testing. Manage data versioning, governance, and lineage using tools such as Unity Catalog or GCP Data Catalog. Partner with cross-functional teams to gather requirements and translate them into technical designs. Provide technical guidance and influence engineering best practices across the team. Contribute to documentation, architectural diagrams, and knowledge sharing., Data engineers are responsible for building reliable and scalable data infrastructure that enables organizations to derive meaningful insights, make data-driven decisions, and unlock the value of their data assets., The role involves leading and managing a team of data engineers, defining and executing the data engineering strategy, and ensuring the effective delivery of data solutions. They provide technical expertise, drive innovation, and collaborate with stakeholders to deliver high quality, scalable, and reliable data infrastructure and solutions.
Requirements
3+ years of experience as a Data Engineer or similar role. Strong hands-on experience with Databricks, including: PySpark/Spark Delta Lake Databricks workflows/jobs Proficiency with GCP: BigQuery Cloud Storage Dataflow or Dataproc Strong coding skills in Python and SQL. Solid understanding of distributed systems, data warehousing, and data architecture principles. Experience with CI/CD tools (GitHub, GitLab, Azure DevOps, or similar)., Databricks or GCP certifications (e.g., Data Engineer, Architect). Experience with Terraform or other Infrastructure-as-Code tools. Knowledge of ML workflows or MLOps frameworks. Familiarity with data governance tools (Unity Catalog, Great Expectations, dbt, etc.).
Benefits & conditions
The base compensation range for this role in the posted location is $76,200 to $187,740
Capgemini provides compensation range information in accordance with applicable national, state, provincial, and local pay transparency laws. The base compensation range listed for this position reflects the minimum and maximum target compensation Capgemini, in good faith, believes it may pay for the role at the time of this posting. This range may be subject to change as permitted by law.
The actual compensation offered to any candidate may fall outside of the posted range and will be determined based on multiple factors legally permitted in the applicable jurisdiction.
These may include, but are not limited to: Geographic location, Education and qualifications, Certifications and licenses, Relevant experience and skills, Seniority and performance, Market and business consideration, Internal pay equity.
It is not typical for candidates to be hired at or near the top of the posted compensation range.
In addition to base salary, this role may be eligible for additional compensation such as variable incentives, bonuses, or commissions, depending on the position and applicable laws.
Capgemini offers a comprehensive, non-negotiable benefits package to all regular, full-time employees. In the U.S. and Canada, available benefits are determined by local policy and eligibility and may include:
-
Paid time off based on employee grade (A-F), defined by policy: Vacation: 12-25 days, depending on grade, Company paid holidays, Personal Days, Sick Leave
-
Medical, dental, and vision coverage (or provincial healthcare coordination in Canada)
-
Retirement savings plans (e.g., 401(k) in the U.S., RRSP in Canada)
-
Life and disability insurance
-
Employee assistance programs
-
Other benefits as provided by local policy and eligibility
Important Notice: Compensation (including bonuses, commissions, or other forms of incentive pay) is not considered earned, vested, or payable until it becomes due under the terms of applicable plans or agreements and is subject to Capgemini's discretion, consistent with applicable laws. The Company reserves the right to amend or withdraw compensation programs at any time, within the limits of applicable legislation.
About the company
Capgemini ist einer der weltweit führenden Anbieter von Management- und IT-Beratung, Technologie-Services und Digitaler Transformation. Als ein Wegbereiter für Innovation unterstützt das Unternehmen seine Kunden bei deren komplexen Herausforderungen rund um Cloud, Digital und Plattformen.