AI & Data Analytics Team

Stellantis
yesterday

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Tech stack

Clean Code Principles
Java
Artificial Intelligence
Airflow
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Data analysis
Azure
Big Data
Cloud Computing
Computer Programming
Databases
Data Architecture
Information Engineering
ETL
Data Transformation
Data Structures
Data Stores
Relational Databases
Database Queries
Software Design Patterns
DevOps
Dimensional Modeling
Distributed Systems
Github
Java Virtual Machine (JVM)
Python
NoSQL
Operational Data Store
Operational Databases
Performance Tuning
Standard Sql
Software Engineering
SQL Databases
Data Streaming
Data Processing
Scripting (Bash/Python/Go/Ruby)
Azure
Spark
Event Driven Architecture
Data Lake
Information Technology
Apache Flink
Luigi
Data Analytics
Google BigQuery
Kafka
Data Management
Api Design
Terraform
Stream Processing
Data Pipelines
Databricks

Job description

The AI & Data Analytics Team is looking for a Senior Data Engineer to join our team. In this role, you will be responsible for designing, building, and optimizing robust data pipelines that process massive datasets in both batch and real-time. You will work at the intersection of software engineering and data science, ensuring that our data architecture is scalable, reliable, and follows industry best practices.

Requirements

  • BA/BSc in Computer Science, Engineering, Mathematics, or a related technical discipline \r\n

  • 5+ years of experience in the data engineering and software development life cycle. \r\n

  • 4+ years of hands-on experience in building and maintaining production data applications, current experience in both relational and columnar data stores. \r\n

  • 4+ years of hands-on experience working with AWS cloud services \r\n

  • Comprehensive experience with one or more programming languages such as Python, Java, or Rust \r\n

  • Comprehensive experience working with Big Data platforms (i.e., Spark, Google Big Query, Azure, AWS S3, etc.) \r\n

  • Familiarity with time series database, data streaming applications, event driven architectures, Kafka, Flink, and more \r\n

  • Experience with workflow management engines (i.e., Airflow, Luigi, Azure Data Factory, etc.) \r\n

  • Experience with designing and implementing real-time pipelines \r\n

  • Experience with data quality and validation \r\n

  • Experience with API design \r\n

  • Distributed Computing: Deep expertise in Apache Spark (Core, SQL, and Structured Streaming). \r\n

  • Programming Mastery: Strong proficiency in Scala or Java. You should be comfortable building production-grade applications in a JVM-based environment. \r\n

  • SQL Proficiency: Advanced knowledge of SQL for data transformation, analysis, and performance tuning. \r\n

  • DevOps & Tools: Hands-on experience with Terraform for infrastructure management and GitHub Actions for CI/CD pipelines. \r\n

  • Software Engineering Foundation: Solid understanding of data structures, algorithms, and design patterns. Experience applying "Clean Code" principles to data engineering. \r\n

  • Stream Processing: Experience with Apache Flink for low-latency stream processing. \r\n

  • Scripting: Proficiency in Python for automation, data analysis, or scripting. \r\n

  • Cloud Platforms: Experience with AWS, Azure, or GCP data services (e.g., EMR, Glue, Databricks). \r\n

  • Data Modeling: Familiarity with dimensional modeling, Lakehouse architectures (Delta Lake, Iceberg), or NoSQL databases. \r\n, r\n \r\n

  • Comprehensive knowledge of relational database concepts, including data architecture, operational data stores, Interface processes, multidimensional modeling, master data management, and data manipulation \r\n

  • Expert knowledge and experience with custom ETL design, implementation and maintenance \r\n

  • Comprehensive experience designing, implementing, and iterating data pipelines using Big Data technologies \r\n

  • Certification in AWS or other cloud providers \r\n

  • Experience with Databricks notebook workflows \r\n

  • Experience with Terraform \r\n

Benefits & conditions

r\n

\r\n

Priorities can change in a fast-paced environment like ours, so this role includes, but is not limited to, the following responsibilities: \r\n

\r\n \r\n

  • Pipeline Development: Design and implement complex data processing pipelines using Apache Spark. \r\n

\r\n \r\n

  • Architectural Leadership: Build scalable, distributed systems that handle high-throughput data streams and large-scale batch processing. \r\n

\r\n \r\n

  • Infrastructure as Code: Manage and provision cloud infrastructure using Terraform. \r\n

\r\n \r\n

  • CI/CD & Automation: Streamline development workflows by implementing and maintaining GitHub Actions for automated testing and deployment. \r\n

\r\n \r\n

  • Code Quality: Uphold rigorous software engineering standards, including comprehensive unit/integration testing, code reviews, and maintainable documentation. \r\n

\r\n \r\n

  • Collaboration: Work closely with stakeholders to translate business requirements into technical specifications. \r\n

Apply for this position