Data Engineer Contract
Role details
Job location
Tech stack
Job description
We are looking for a forward-thinking Data Engineer with strong proficiency in both SQL and Python to join our Group Operations Data Transformation Team. The successful candidate will be responsible for supporting and enhancing our Microsoft Fabric platform, developing data pipelines, and crucially, identifying and resolving data quality issues at source.
Unlike a standard developer role, this position has a critical focus on Data Quality and Governance . You will not just be moving data; you will be responsible for ensuring its integrity. You will be tasked with identifying quality issues at the source (e.g., detecting null anomalies or schema drifts) and designing automated, AI-driven processes to rectify them.
Working closely with Analysts, BI Developers, and Information Systems (IS), you will transform raw system data into a "Gold Standard" format to enable regular reporting and insight. You will be expected to understand the business context behind the data, ensuring that what we build drives tangible value for our clients., * Design, develop, and maintain robust ETL/ELT pipelines using Microsoft Fabric (primary), Azure Data Factory, and Python.
- Manage the ingestion of raw data from diverse sources (CAFM, Finance Systems, Salesforce, IoT, HR) into the Data Lakehouse.
- Work with Cloudera and Azure stack technologies to deliver comprehensive data solutions Collaborate with the architecture team to design the optimal, future-proofed solution and support the transition from the 'as-is' to the 'to-be' solution.
Data Quality (DQ) & Governance
-
Proactive DQ Monitoring: Implement automated scripts (SQL/Python) to identify data quality issues at the source
-
AI & Automation: Design, implement, and maintain processes to improve data quality using AI and Automation
-
Governance Framework: Maintain Data Governance standards, ensuring data dictionaries, lineage, and schematics are kept up to date.
-
Conduct root cause analysis on data discrepancies and collaborate with source system owners to rectify issues upstream.
Collaboration & Optimisation
- Assemble large, complex datasets that meet functional and non-functional business requirements.
- Review and improve user-generated scripts (SQL/Python) for performance and scalability.
- Conduct performance analysis and provide recommendations for optimisation .
- Communicate technical developments and infrastructure changes clearly to non-technical stakeholders and business users.
Requirements
- SQL & Python: Expert-level SQL skills for complex querying and stored procedures, combined with strong Python skills for data manipulation (Pandas, PySpark) and automation.
- Microsoft Fabric & Azure: Experience with the Azure Data Stack (ADF, Synapse, Data Lake) is essential. Experience with or a strong desire to master Microsoft Fabric is required.
- Big Data Ecosystems: Knowledge of Cloudera/Hadoop environments is highly beneficial.
- Data Quality Automation: Demonstrated experience in automating data quality checks and using logic/AI to improve data integrity., * 3+ years of experience in a Data Engineer role, who has attained a Graduate degree in Computer Science, Statistics, Informatics, Information Systems or another quantitative field and is keen to progress their career in Data Engineering and Data Development
- Proven track record of delivering ETL/ELT projects (ingestion, modelling , mapping).
- Experience implementing Data Governance controls and documentation.
Soft Skills
- Investigative Mindset: You don't just fix the error; you analyse why the error happened.
- Communication: Excellent written and verbal skills; able to translate "data issues" into "business impacts".
- Self-Starter: Motivated and organised, with an ability to work to tight deadlines in a hybrid environment.
- Remote Working: A good, reliable internet connection at home is required.