Data Engineer
Techstra Solutions
Pittsburgh, United States of America
1 month ago
Role details
Contract type
Permanent contract Employment type
Full-time (> 32 hours) Working hours
Regular working hours Languages
English Experience level
SeniorJob location
Pittsburgh, United States of America
Tech stack
Airflow
Azure
Big Data
Software as a Service
Cloud Computing
Profiling
Code Review
Databases
Continuous Integration
Data as a Services
Data Validation
Data Governance
Data Integration
ETL
Data Profiling
Data Sharing
DevOps
Hadoop
Python
Microsoft Office
Microsoft SQL Server
MySQL
Oracle Applications
Platform as a Service (PAAS)
Power BI
Kusto Query Language
Shell Script
SQL Databases
Data Streaming
Systems Integration
Tableau
Google Cloud Platform
Data Storage Technologies
Data Ingestion
Flask
Spark
GIT
FastAPI
Pandas
Data Lake
PySpark
Information Technology
Real Time Data
Kafka
Wikis
Data Management
Tools for Reporting
Cloud Integration
Api Design
REST
Oracle Cloud Infrastructure
Data Pipelines
Docker
Databricks
Job description
We are seeking an experienced Data Engineer contractor to support our operations. This individual will design, build, and optimize data pipelines and infrastructure, enabling advanced analytics, process automation, and data-driven decision-making. The Data Engineer will work closely with data scientist, process engineering, and IT teams to ensure data reliability and actionable insights across the manufacturing lifecycle., * Develop/maintain scalable and reliable data pipelines for industrial data (like real-time streaming, time series, IoT, sensors, MES, ERP systems data)
- Integrate data from different sources (databases, clouds, on premises) and Engineer workflows for efficient ETL/ELT processing and data validation.
- Collaborate with architects, data engineers, data scientists, analysts, and business stakeholders to define and deliver solutions.
- Collaborate with IT admins, network/security engineers, and cross-functional teams to support stable production operations and troubleshoot infrastructure issues (including managing and integrating IaaC, PaaS, and SaaS solutions).
- Capable of managing backlog, supporting QA/testing, and communicating requirements with business stakeholders in the steel manufacturing domain.
- Mentoring team members, providing guidance, facilitating skill growth, offering technical coaching, and encouraging best practices across teams via code reviews.
- Build and maintain data infrastructure in compliance with data governance and security best practices
Requirements
- Bachelor's degree in computer science or related fields with 5+ years' experience as a Data Engineer.
- Strong experience in building, maintaining, and optimizing ETL/ELT data pipelines using Python, Pandas, PySpark and orchestrating workflows like Apache Airflow and Kedro framework.
- Advanced SQL/ KQL query development and optimization across Oracle, MSSQL, and MySQL databases (hosted on-premises or via PaaS offerings).
- Developing and consuming Flask-based and Fast API RESTful APIs for data services and integration.
- Proficiency in Linux shell scripting for automation and data workflow management.
- Experience with DevOps practices, including CI/CD for data pipelines and use of tools such as Git, Docker, and IaaC frameworks for provisioning and deployment.
- Hands-on experience deploying solutions across multiple clouds (OCI, Azure, Google Cloud Platform), including the setup of cross-cloud data integration and transfer techniques.
- Experience with cloud platforms (OCI, Azure, Google) and big data tools (Spark, Hadoop, Kafka, Databricks)
- Understanding data modeling, data profiling, data quality, data lake/warehouse architectures, and data indigestion from operational technologies.
- Familiarity with industrial protocols, time-series databases (like OSIsoft PI), and manufacturing data (MES, PLC)
- Strong troubleshooting, process automation, and root-cause analysis skills
Preferred Qualifications
- Role Responsibility Area Preferred Tools & Skills
- Data Ingestion Pipeline Python, PySpark, Airflow, Kedro, Linux shell scripting
- API Development Flask, Fast API, RESTful design
- Data Storage & Querying SQL (Oracle, MSSQL, MySQL), KQL (Azure Data Explorer), Bigdata (Hadoop, Oracle BDS), OSIsoft PI
- Cloud Integration Multi-cloud platforms (OCI, Azure, Google Cloud Platform), Data Sharing across cloud (Data Bricks)
- Real-Time Data Streaming Kafka, Azure Event Hub, EMQX
- Reporting Tools Tableau, OAC, Power BI
- Collaboration Wiki, Azure DevOps Boards, MS Office 365
- Data Governance & Quality Data profiling/validation tools (Pandas Profiling), SaaS monitoring (e.g., Great Expectations), lineage tracking (Cloud Data Catalogs)
About the company
Techstra Solutions is a certified woman-owned (WBENC) management consulting firm specializing in strategy, technology, and implementation services for large organizations undergoing digital and talent transformation. Our experienced team partners with clients to co-create innovative solutions in applications, data, AI, and automation that accelerate measurable, sustainable change. From advisory consulting through technical execution, we are dedicated to driving world-class business solutions that fit your strategic requirements and deliver results. For more information: [INS: :INS]