Cloud Data Engineer

NTT DATA, Inc.
Plano, United States of America
yesterday

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Plano, United States of America

Tech stack

API
Azure
Batch Processing
Big Data
Cloud Database
Continuous Integration
Data Governance
Data Infrastructure
ETL
Data Systems
Data Warehousing
DevOps
Revision Control Systems
Hive
Python
SQL Azure
Performance Tuning
Power BI
Azure
Standard Sql
Scala
SQL Databases
Data Streaming
Teradata
Data Logging
Data Processing
Data Ingestion
Microsoft Power Automate
Azure
Spark
GIT
Pandas
Data Lake
PySpark
Information Technology
Optimization Algorithms
Deployment Automation
Star Schema
Kafka
Cosmos DB
Data Management
Tools for Reporting
Video Streaming
Azure
Data Pipelines
Serverless Computing
ServiceNow
Databricks
Control M

Job description

  • We are looking for a skilled Cloud Data Engineer with strong experience in Azure-based data platforms, particularly Azure Data Factory, Databricks and Python. The ideal candidate will design, build, and optimize scalable data pipelines and data solutions to support analytics and business intelligence needs. Experience with Teradata ETL development is a strong plus.

Day to Day job Duties: (what this person will do on a daily/weekly basis)

  • As a Cloud Data Engineer in the Service Delivery Tier 1 Support & Development team, you will be responsible for designing, developing, and maintaining robust data pipelines using Azure Data Factory (ADF). Build and optimize data processing workflows using Azure Databricks (PySpark/Scala).
  • Develop scalable data architectures in Azure (Data Lake, Synapse, etc.)
  • Develop and maintain data solutions using Python for data transformation and automation.
  • Perform data ingestion, transformation, and integration from various sources (structured/unstructured)
  • Ensure data quality, integrity, and governance across pipelines
  • Optimize performance and cost of data processing solutions
  • Collaborate with data analysts, data scientists, and business stakeholders to deliver data solutions
  • Implement monitoring, logging, and alerting data pipelines
  • Support CI/CD processes and deployment automation for data workflows
  • Monitoring and responding to the issues or failures in the Cloud Data Engineering processes for data processing and loading. Data loading and processing occur as either API, Streaming or Batch processing. These data processes are critical to retain data freshness and continuity across a large Microsoft Azure EDL. The Microsoft Azure data sources are critical for business reporting, critical business decisions and revenue generating operational processes.
  • Be a part of a cross-North America Cloud Data Engineering Delivery team that will partner together making a difference every day for the largest global rental car company.
  • Having foundational skills and experience in a data warehouse star schema with ETL concepts. Previous SQL knowledge and experience with a large EDL environment.

Requirements

Basic Qualifications: (what are the skills required to this job with minimum years of experience on each)

  • Must have 8 - 10 years' experience in Cloud Data Engineering Development & support roles preferably in Microsoft Azure Data Warehouse.
  • Must have Cloud Data Engineering Support experience working with large datasets and complex data environments, processes, and associated solutions.
  • Knowledge and experience with data warehouse star schema and ETL concepts
  • Proficiency in SQL and data modeling concepts
  • 8-10 years' experience with Databricks (workflows and jobs orchestrations), Py Spark, Python, Data factory, Delta Optimization techniques, SQL, transformations, Databricks Cluster
  • Knowledge of Streaming Support with Event Hub, Cosmos DB, Scala, Azure SQL DB and Azure App Insights, key vaults, Azure Functions and logic apps, Azure AAD + key vaults, spark SQL, pandas etc.
  • Knowledge of automated and scheduled batch job tools such as IBM IWS and Control-M.
  • Knowledge of data governance and unity catalog.
  • Knowledge of performance tuning and optimization techniques
  • Familiarity with version control tools (e.g., Git)
  • Knowledge Service now, incident management, ticket handling, change management, on call roaster management, etc.
  • Must have good oral and written communication skills to effectively communicate with various IT teams and business representatives.
  • Must be able to both collaborate in a team-oriented environment and work independently with direction.

Nice to Have; (But not a must)

  • Experience in ETL development using Teradata
  • Knowledge of Azure Synapse Analytics
  • Experience with streaming technologies (e.g., Kafka, Event Hub)
  • Exposure to DevOps practices and tools (Azure DevOps, CI/CD pipelines)
  • Understanding of data governance and security best practices
  • Experience with Power BI or other reporting tools

Education

  • Bachelor's degree in computer science, Engineering, or a related field (or equivalent experience)

About the company

NTT DATA is a $30 billion trusted global innovator of business and technology services. We serve 75% of the Fortune Global 100 and are committed to helping clients innovate, optimize and transform for long term success. As a Global Top Employer, we have diverse experts in more than 50 countries and a robust partner ecosystem of established and start-up companies. Our services include business and technology consulting, data and artificial intelligence, industry solutions, as well as the development, implementation and management of applications, infrastructure and connectivity. We are one of the leading providers of digital and AI infrastructure in the world. NTT DATA is a part of NTT Group, which invests over $3.6 billion each year in R&D to help organizations and society move confidently and sustainably into the digital future. Visit us at us.nttdata.com

Apply for this position