Data Engineer

Guidehouse Inc.
San Antonio, United States of America
2 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Intermediate

Job location

San Antonio, United States of America

Tech stack

Azure
Data as a Services
ETL
Python
Performance Tuning
Azure
Scala
SQL Databases
Unstructured Data
Data Processing
Azure
Snowflake
Spark
Data Lake
Information Technology
Azure
Data Pipelines
Databricks

Job description

  • Design, build, and operate Azure Lakehouse architectures using Azure Databricks, Azure Data Lake Storage (ADLS Gen2), Azure Synapse Analytics, and Azure Data Factory to support analytical and operational workloads.
  • Process large-scale structured and unstructured datasets using optimized batch and streaming pipelines leveraging Apache Spark, Delta Lake, Python, SQL, and Scala.
  • Design, develop, and maintain scalable ETL/ELT pipelines using Databricks Workflows, Spark jobs, and Delta Lake, ensuring reliability, performance, and data quality at enterprise scale.
  • Implement real-time and batch data processing solutions and optimize pipelines for production use.

Requirements

  • US Citizenship required.
  • Must be able to OBTAIN and MAINTAIN a Federal or DoD "PUBLIC TRUST" security clearance.
  • Bachelor's degree obtained in Computer Science, Engineering, Data Science, or equivalent experience.
  • 5+ years of experience as a Data Engineer.
  • 3+ years delivering enterprise solutions on Azure, specifically using Azure Databricks and Azure-native data services.
  • Hands-on experience designing, building, and operating Azure Lakehouse architectures using Azure Databricks, ADLS Gen2, Azure Synapse Analytics, and Azure Data Factory.
  • Deep expertise in Apache Spark, Delta Lake, Python, SQL, and Scala, with demonstrated ability to process large-scale structured and unstructured datasets using optimized batch and streaming pipelines.
  • Proven experience designing, developing, and maintaining scalable ETL/ELT pipelines using Databricks Workflows, Spark jobs, and Delta Lake, ensuring reliability, performance, and data quality.

What Would Be Nice To Have:

  • Experience implementing real-time and batch streaming solutions.
  • Master's degree in computer science, Engineering, Data Science, or equivalent experience
  • Experience with Snowflake-specific capabilities such as Snowpipe, Snowpark, Streams & Tasks, Snowflake performance tuning, or Snowflake certification.

Benefits & conditions

Guidehouse offers a comprehensive, total rewards package that includes competitive compensation and a flexible benefits package that reflects our commitment to creating a diverse and supportive workplace.

Benefits include:

  • Medical, Rx, Dental & Vision Insurance
  • Personal and Family Sick Time & Company Paid Holidays
  • Position may be eligible for a discretionary variable incentive bonus
  • Parental Leave and Adoption Assistance
  • 401(k) Retirement Plan
  • Basic Life & Supplemental Life
  • Health Savings Account, Dental/Vision & Dependent Care Flexible Spending Accounts
  • Short-Term & Long-Term Disability
  • Student Loan PayDown
  • Tuition Reimbursement, Personal Development & Learning Opportunities
  • Skills Development & Certifications
  • Employee Referral Program
  • Corporate Sponsored Events & Community Outreach
  • Emergency Back-Up Childcare Program
  • Mobility Stipend

About Guidehouse

Guidehouse is an Equal Opportunity Employer-Protected Veterans, Individuals with Disabilities or any other basis protected by law, ordinance, or regulation.

Guidehouse will consider for employment qualified applicants with criminal histories in a manner consistent with the requirements of applicable law or ordinance including the Fair Chance Ordinance of Los Angeles and San Francisco.

Apply for this position