Senior Data Engineer (AWS)

Provectus
3 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Remote

Tech stack

API
Artificial Intelligence
Airflow
Amazon Web Services (AWS)
Apache HTTP Server
Azure
Big Data
Cloud Computing
Information Engineering
Data Governance
ETL
Data Systems
Data Warehousing
DevOps
Python
Machine Learning
Power BI
Software Engineering
SQL Databases
Data Streaming
Tableau
Google Cloud Platform
Chatbots
Flask
Snowflake
Spark
Generative AI
Cloudformation
FastAPI
Druid
Kafka
Data Management
Terraform
Looker Analytics
Data Pipelines
Databricks
Programming Languages

Job description

Provectus helps companies adopt ML/AI to transform the ways they operate, compete, and drive value. The focus of the company is on building ML Infrastructure to drive end-to-end AI transformations, assisting businesses in adopting the right AI use cases, and scaling their AI initiatives organization-wide in such industries as Healthcare & Life Sciences, Retail & CPG, Media & Entertainment, Manufacturing, and Internet businesses., * Collaborate closely with clients to deeply understand their existing IT environments, applications, business requirements, and digital transformation goals;

  • Collect and manage large volumes of varied data sets;
  • Work directly with ML Engineers to create robust and resilient data pipelines that feed Data Products;
  • Define data models that integrate disparate data across the organization;
  • Design, implement, and maintain ETL/ELT data pipelines;
  • Perform data transformations using tools such as Spark, Trino, and AWS Athena to handle large volumes of data efficiently;
  • Develop, continuously test, and deploy Data API Products with Python and frameworks like Flask or FastAPI.

Requirements

Do you have experience in Terraform?, Do you have a Master's degree?, We are seeking a talented and experienced Data Engineer to join our team at Provectus. As part of our diverse practices, including Data, Machine Learning, DevOps, Application Development, and QA, you will collaborate with a multidisciplinary team of data engineers, machine learning engineers, and application developers. You will encounter numerous technical challenges and will have the opportunity to contribute to the internal solutions, engage in R&D activities, providing an excellent environment for professional growth., * 5+ years of experience in data engineering;

  • Experience in AWS;
  • Experience handling real-time and batch data flow and data warehousing with tools and technologies like Airflow, Dagster, Kafka, Apache Druid, Spark, dbt, etc.;
  • Proficiency in programming languages relevant to data engineering, such as Python and SQL;
  • Proficiency with Infrastructure as Code (IaC) technologies like Terraform or AWS CloudFormation;
  • Experience in building scalable APIs;
  • Familiarity with Data Governance aspects like Quality, Discovery, Lineage, Security, Business Glossary, Modeling, Master Data, and Cost Optimization;
  • Upper-Intermediate or higher English skills;
  • Ability to take ownership, solve problems proactively, and collaborate effectively in dynamic settings., * Experience with Cloud Data Platforms (e.g., Snowflake, Databricks);
  • Experience in building Generative AI Applications (e.g., chatbots, RAG systems);
  • Relevant AWS, GCP, Azure, Databricks certifications;
  • Knowledge of BI Tools (Power BI, QuickSight, Looker, Tableau, etc.);
  • Experience in building Data Solutions in a Data Mesh architecture.

Apply for this position