Senior Data Engineer

Hawk
22 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Tech stack

Airflow
Amazon Web Services (AWS)
Azure
Databases
Data Architecture
Information Engineering
Data Infrastructure
Data Systems
Distributed Data Store
Elasticsearch
Graph Database
Python
PostgreSQL
Machine Learning
Neo4j
Query Optimization
Power BI
Tableau
Google Cloud Platform
Data Storage Technologies
Spark
Data Lake
Information Technology
Apache Flink
Kafka
Presto
Tools for Reporting
Stream Processing
Data Pipelines
Databricks

Job description

  • Build and maintain scalable data infrastructure: Design, implement, and optimize our datalake and associated pipelines to support reporting, analytics, and machine learning workloads.
  • Shape architecture and tooling: Lead discussions on data architecture, recommend tools and frameworks, and ensure the adoption of best practices for distributed data processing and orchestration.
  • Data preparation and quality: Prepare datasets for internal and external business reporting, ensuring data quality, consistency, and accessibility.
  • Machine learning readiness: Collaborate with data scientists to prepare and optimize datasets for machine learning and advanced analytics.
  • Collaborate across teams: Work closely with operations, data science, and business stakeholders to understand requirements and deliver solutions that align with their needs.
  • Drive innovation: Explore and implement new technologies and methods to optimize data storage, processing, and access.

Requirements

Do you have experience in Tableau?, Do you have a Master's degree?, + Bachelor's or master's degree in computer science or a related technical field.

  • Technical expertise:
  • Strong expertise in building data pipelines to support business critical reporting infrastructure.
  • Strong experience with cloud platforms (AWS, GCP, Azure).
  • Proficient in distributed data processing tools (e.g., Spark) and stream processing frameworks (e.g., Kafka, Flink).
  • Strong understanding of lake house architectures and supporting technologies (e.g., Delta Lake, Iceberg, Hudi), query optimization tools (e.g., Trino, Presto), and orchestration tools (e.g., Airflow)
  • Experience with database technologies (Elasticsearch, PostgreSQL). Knowledge of graph databases (e.g. Neo4j) is a plus.
  • Familiarity with reporting tools (e.g., Tableau, Power BI, Databricks Dashboards).
  • Advanced Python programming skills.
  • Hands-on experience with distributed machine learning frameworks and pipelines is a plus.
  • Proven experience:
  • 5+ years in data engineering or a related role, with a track record of delivering business value through scalable data solutions.
  • Collaborative mindset:
  • Strong interpersonal and communication skills, enabling you to collaborate effectively across diverse teams.
  • Commitment to quality:
  • Passion for building robust, high-quality solutions with a focus on innovation and continuous improvement.

About the company

Hawk is the leading provider of AI-supported anti-money laundering and fraud detection technology. Banks and payment providers globally are using Hawk's powerful combination of traditional rules and explainable AI to improve the effectiveness of their AML compliance and fraud prevention by identifying more crime while maximizing efficiency by reducing false positives. With our solution, we are playing a vital role in the global fight against Money Laundering, Fraud, or the financing of terrorism. We offer a culture of mutual trust, support and passion - while providing individuals with opportunities to grow professionally and make a difference in the world.

Apply for this position