Data Engineer
Role details
Job location
Tech stack
Job description
Data Engineer Long Stratton, Norwich, Norfolk £56,000 per annum Full Time: 37hrs per week Saffron is looking for a talented Data Engineer to help drive the next stage of our data transformation. This role is all about building and optimising our Azure-based data platform, developing high-performing pipelines in Azure Data Factory, and supporting our move toward Microsoft Fabric. You will work closely with BI Analysts and teams across the business to deliver reliable, high-quality data that powers smarter decisions and sharper insights. It is a chance to shape a modern, scalable data environment and make a real impact on how we use data across the organisation. Key Responsibilities: * Design, build, and maintain a scalable Azure-based data warehouse that meets the current and future requirements of the Data & Analytics team. * Lead the introduction, adoption, and optimisation of Microsoft Fabric (e.g., Lakehouse, Warehouse, Data Engineering, Pipelines). * Apply CI/CD practices
Requirements
(e.g., Azure DevOps) for version control, deployment automation, and environment management. * Implement data quality checks, pipeline observability, alerting, and automated monitoring to ensure consistent platform reliability. * Work collaboratively with data owners and the wider data team to ensure data definitions, lineage, and ownership are clearly established. * Work collaboratively with data owners and the wider data team to ensure data definitions, lineage, and ownership are clearly established. * Provide technical guidance and coaching to the wider data team members on data engineering best practices. For a full list of responsibilities please see the attached Role Profile Our Ideal Candidate Will Have: Education and Qualifications: * Degree in Computer Science, Data Engineering, Mathematics, or a related discipline, or equivalent experience (E) * Microsoft certifications in SQL, Fabric, including Power BI, or other Azure Data Services (D) Experience: * Advanced SQL skills, including optimisation of complex queries (E). * Experience building data pipelines and ETL/ELT workflows using tools such as: * Azure Data Factory, Databricks, Airflow, Luigi, or similar (E) * Strong understanding of data modelling (E) * Programming skills in Python and/or Scala for data processing (D). * Experience with machine learning pipelines or MLOps frameworks (D). Personal Attributes: * Confident communicator able to engage both technical and non-technical audiences. * Proactive, innovative, and committed to continuous improvement. * Collaborative, with mentoring and leadership capabilities. * Customer-focused, with a commitment to improving services through data. * Experience working in a busy, fast-paced workload, and managing multiple projects to meet deadlines