Data Engineer (Developer)
Engineergazelle Global Consulting Ltd
Charing Cross, United Kingdom
3 days ago
Role details
Contract type
Permanent contract Employment type
Part-time (≤ 32 hours) Working hours
Regular working hours Languages
English Experience level
SeniorJob location
Charing Cross, United Kingdom
Tech stack
Java
Agile Methodologies
Unit Testing
Azure
Batch Processing
Computer Programming
Data Validation
ETL
Data Masking
Hive
Python
NoSQL
Object-Oriented Software Development
Data Processing
Delivery Pipeline
Spark
Gitlab
Microsoft Fabric
PySpark
Spark Streaming
Data Pipelines
Job description
Design, build, and optimise scalable data pipelines for both batch and streaming workloads:
- Develop dataflows and semantic models aligned to analytics and reporting needs
- Implement complex transformations and performance-focused data processing logic
- Apply data validation, cleansing, and profiling techniques to ensure accuracy and consistency
- Implement access controls, data masking, and compliance-aligned security protocols
- Tune workloads and optimise performance across Spark, Fabric, and Azure components
- Translate business requirements into technical solutions through close collaboration with analysts and stakeholders
- Maintain clear documentation and contribute to internal knowledge repositories
Requirements
Strong experience developing within Microsoft Azure and Microsoft Fabric:
- Proficiency in Spark programming including DataFrames, RDDs, and Spark SQL
- Python and PySpark development experience, including notebook-based workflows
- Hands-on experience with Spark streaming and batch processing
- Delta table optimisation and Fabric Spark job development
- Solid Java programming and OOP understanding
- Experience working with relational and NoSQL databases
- Familiarity with GitLab, unit testing, and CI/CD pipelines
- Strong troubleshooting ability and experience working in Agile environments
- Excellent communication skills with stakeholder-facing experience
- Practical experience building ETL workflows, lakehouse architectures, dataflows, and semantic models
- Exposure to time-series data, financial market feeds, transactional records, and risk-related datasets
About the company
We are supporting a leading global financial markets infrastructure and data provider as they modernise and scale their core data engineering capabilities. This role sits at the centre of their transformation programme, delivering high-quality data pipelines, models, and platforms that underpin critical services across the business.