Data Analytics Principal Software Engineer

Lexstra Plc
Richmond, United Kingdom
2 days ago

Role details

Contract type
Temporary contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Compensation
£ 163K

Job location

Richmond, United Kingdom

Tech stack

Artificial Intelligence
Amazon Web Services (AWS)
Apache HTTP Server
Batch Processing
Google BigQuery
C Sharp (Programming Language)
Cloud Computing
Cloud Storage
Computer Programming
Data as a Services
Data Architecture
Information Engineering
Data Governance
ETL
Data Systems
DevOps
Disaster Recovery
Fault Tolerance
MapReduce
Identity and Access Management
Python
Machine Learning
Online Analytical Processing
Online Transaction Processing
Role-Based Access Control
Power BI
Mixpanel
Azure
SQL Databases
Data Streaming
Scripting (Bash/Python/Go/Ruby)
Delivery Pipeline
Large Language Models
Spark
Cloudformation
Data Lake
Collibra
Amazon Web Services (AWS)
Data Analytics
Kafka
Data Management
Amazon Web Services (AWS)
Terraform
Azure
Databricks

Requirements

You will need to have software developer background with hands-on experience on several large enterprise data lake projects, preferably with strong Python. Must also have experience as a technical lead across multiple teams (both onshore and offshore) building data platforms, customer facing data products and/or machine learning systems, with experience of product analytics tools (Mixpanel, Power BI, Athena). Experience working with LLMs in Data engineering and using AI as an accelerator is also key to this role.

Technology Requirements:

Data Architecture & Design: Data Lakes (eg, AWS S3, Azure Data Lake, Google Cloud Storage), Data Mesh principles, domain-oriented data ownership and federated governance, data modelling (OLAP/OLTP, dimensional modelling, schema evolution)

Data Engineering & Pipelines: ETL pipelines (using tools like AWS Glue, Apache Spark), Map-Reduce, streaming data platforms (eg, Kafka, SQS), Real Time and batch processing paradigms

Cloud & Infrastructure: cloud-native data services (AWS Glue, Azure Synapse, GCP BigQuery, Databricks), Infrastructure-as-Code (IaC) (using Terraform, CloudFormation, Lakeformation)

Programming & Scripting: Python and SQL, C#, CI/CD pipelines and DevOps practices for data workflows

Data Governance & Security: Data cataloging and lineage tools (eg, Collibra, Apache Atlas, OpenMetaData), data privacy, encryption, access control (eg, IAM, RBAC, ABAC), and compliance frameworks (GDPR)

Observability & Reliability: Monitoring and alerting for data systems, data quality frameworks (eg, Great Expectations, Monte Carlo), designing for resilience, fault tolerance, and disaster recovery

Apply for this position