DevOps Engineer

Espire Infolabs Limited
Charing Cross, United Kingdom
5 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Charing Cross, United Kingdom

Tech stack

Amazon Web Services (AWS)
Data analysis
Azure
Software Quality
Code Review
Data Governance
Data Infrastructure
Data Integration
DevOps
R
Python
Software Deployment
Enterprise Data Management
Jupyter Notebook
Parquet
Cloud Platform System
GIT
Git Flow
Machine Learning Operations
Data Pipelines
Docker

Job description

We're looking for a senior, hands-on Data Platform & DevOps Engineer to act as the critical link between Data Science and Engineering/Operations. This is an exciting opportunity to shape how data science workflows are built, standardised, and scaled - taking experiments from the notebook all the way through to reliable, automated production deployments.

You'll collaborate closely with data scientists, engineers, and platform teams, bringing together deep analytical tooling knowledge and modern DevOps practices to build a robust, production-ready data science environment. What you'll be doing

  • Serve as the technical bridge between Data Science and Engineering/Operations, driving collaboration and alignment across teams
  • Enable and optimise analytics workflows using Python, R, Jupyter Notebooks/JupyterLab, and analytical data formats such as Parquet
  • Own and evolve a multi-user JupyterLab environment, making it a best-in-class collaborative space for data scientists
  • Extend and maintain an existing Python framework for secure, enterprise-grade data integrations
  • Lead advanced Git practices across the organisation - branching strategies, merging, conflict resolution, code reviews, and repository standards
  • Design, build, and maintain CI/CD pipelines that support the full build-test-deploy lifecycle for analytics workloads
  • Containerise data science environments using Docker, ensuring portability and reproducibility at scale
  • Partner with cloud and infrastructure teams to deploy and operate analytics workloads in cloud environments
  • Champion engineering best practices around code quality, security, scalability, and operational stability

Requirements

Essential

  • Strong, demonstrable Python experience - this is a core requirement; working knowledge of R is also expected
  • Extensive hands-on experience with Jupyter Notebooks and JupyterLab in collaborative or multi-user settings
  • Solid understanding of analytical data formats, particularly Parquet
  • Advanced Git proficiency - branching, merging, conflict resolution, and enforcing team-wide standards
  • Proven track record designing and maintaining CI/CD pipelines for data or analytics workloads
  • Practical experience with Docker and containerised environments
  • A background that spans both data science and engineering/operations - you're comfortable in both worlds
  • Strong communication skills and a collaborative, solutions-focused mindset

Desirable

  • Experience working with enterprise data platforms or large-scale analytics ecosystems
  • Familiarity with one or more major cloud providers - AWS, Azure, or GCP
  • A history of industrialising data science or machine learning workflows
  • Knowledge of data governance, access controls, and compliance frameworks

Apply for this position