DevOps Engineer
Espire Infolabs Limited
Charing Cross, United Kingdom
5 days ago
Role details
Contract type
Permanent contract Employment type
Full-time (> 32 hours) Working hours
Regular working hours Languages
English Experience level
SeniorJob location
Charing Cross, United Kingdom
Tech stack
Amazon Web Services (AWS)
Data analysis
Azure
Software Quality
Code Review
Data Governance
Data Infrastructure
Data Integration
DevOps
R
Python
Software Deployment
Enterprise Data Management
Jupyter Notebook
Parquet
Cloud Platform System
GIT
Git Flow
Machine Learning Operations
Data Pipelines
Docker
Job description
We're looking for a senior, hands-on Data Platform & DevOps Engineer to act as the critical link between Data Science and Engineering/Operations. This is an exciting opportunity to shape how data science workflows are built, standardised, and scaled - taking experiments from the notebook all the way through to reliable, automated production deployments.
You'll collaborate closely with data scientists, engineers, and platform teams, bringing together deep analytical tooling knowledge and modern DevOps practices to build a robust, production-ready data science environment. What you'll be doing
- Serve as the technical bridge between Data Science and Engineering/Operations, driving collaboration and alignment across teams
- Enable and optimise analytics workflows using Python, R, Jupyter Notebooks/JupyterLab, and analytical data formats such as Parquet
- Own and evolve a multi-user JupyterLab environment, making it a best-in-class collaborative space for data scientists
- Extend and maintain an existing Python framework for secure, enterprise-grade data integrations
- Lead advanced Git practices across the organisation - branching strategies, merging, conflict resolution, code reviews, and repository standards
- Design, build, and maintain CI/CD pipelines that support the full build-test-deploy lifecycle for analytics workloads
- Containerise data science environments using Docker, ensuring portability and reproducibility at scale
- Partner with cloud and infrastructure teams to deploy and operate analytics workloads in cloud environments
- Champion engineering best practices around code quality, security, scalability, and operational stability
Requirements
Essential
- Strong, demonstrable Python experience - this is a core requirement; working knowledge of R is also expected
- Extensive hands-on experience with Jupyter Notebooks and JupyterLab in collaborative or multi-user settings
- Solid understanding of analytical data formats, particularly Parquet
- Advanced Git proficiency - branching, merging, conflict resolution, and enforcing team-wide standards
- Proven track record designing and maintaining CI/CD pipelines for data or analytics workloads
- Practical experience with Docker and containerised environments
- A background that spans both data science and engineering/operations - you're comfortable in both worlds
- Strong communication skills and a collaborative, solutions-focused mindset
Desirable
- Experience working with enterprise data platforms or large-scale analytics ecosystems
- Familiarity with one or more major cloud providers - AWS, Azure, or GCP
- A history of industrialising data science or machine learning workflows
- Knowledge of data governance, access controls, and compliance frameworks