Data Engineer III
Role details
Job location
Tech stack
Job description
As a Data Engineer at JPMorgan Chase within Personal Investing, you will build and operate a robust cloud-native data platform and pipelines that power analytics, regulatory reporting, and data-promoten applications at scale. You will help us deliver reliable, scalable, observable, and secure data solutions across cloud-native services, lakehouse architectures, data warehousing, and streaming systems. You'll partner with teammates to build consistent, maintainable pipelines and contribute across the software delivery lifecycle from requirements through support., * Build and maintain scalable, reusable data processing and data quality frameworks using Python, PySpark, and dbt
- Build and operate batch and streaming data pipelines with strong scalability, performance, and fault tolerance
- Develop and manage workflow orchestration using tools such as Apache Airflow to support reliable, observable, and well-scheduled data movement and transformations
- Implement and optimize data models and warehouse structures to support analytics and business intelligence workloads
- Write clean, testable Python/PySpark code using object-oriented principles and unit testing
- Implement infrastructure-as-code for the data platform using Terraform
- Containerize and deploy services using Docker, Kubernetes, and Helm
- Contribute across the software development lifecycle, including requirements, design, development, testing, deployment, release, and support
- Collaborate with teammates in an agile, dynamic environment to deliver reliable outcomes
Requirements
- Degree in Computer Science or a STEM-related field (or equivalent)
- Experience working in an agile and dynamic environment
- Experience across the software development lifecycle (requirements, design, architecture, development, testing, deployment, release, and support)
- At least 5 years of recent, hands-on professional experience actively coding as a data engineer
- Hands-on experience with major cloud technologies (e.g., AWS, Google Cloud, or Azure)
- Experience writing Python using object-oriented programming and unit/integration testing practices
- Experience with SQL and familiarity with SQL-based workflow management tools such as dbt
- Experience with orchestration tools such as Airflow (or similar)
- Understanding of messaging/streaming systems such as Kafka or Pub/Sub (or similar)
- Familiarity with infrastructure-as-code (e.g., Terraform) for cloud-based data infrastructure, * Data modeling skills
- Experience with data streaming and scalable processing frameworks (e.g., Spark, Flink, Beam, or similar)
- Experience automating deployment, releases, and testing in continuous integration and continuous delivery pipelines
- Experience with lakehouse patterns and table formats (e.g., Apache Iceberg)
- Experience with federated query engines such as Trino
- Experience designing automated tests (unit, component, integration, and end-to-end), including use of mocking frameworks
- Experience with containers and container-based deployment environments (e.g., Docker, Kubernetes, or similar)