Senior Data Platform / DevOps Engineer (Agile, Digital Platforms)

Nextlink Solutions
2 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior

Job location

Tech stack

Java
Agile Methodologies
Amazon Web Services (AWS)
Apache HTTP Server
Big Data
Cloud Computing
Computer Programming
ETL
Linux
DevOps
Distributed Systems
Jinja (Template Engine)
Python
Octopus Deploy
Scrum
Ansible
Data Streaming
Web Platforms
Parquet
Scripting (Bash/Python/Go/Ruby)
Spark
Firewalls (Computer Science)
Data Lake
Kubernetes
Collibra
Kafka
Data Management
Cloud Migration
Dataiku
Puppet
Data Pipelines
Devsecops
Docker
Jenkins
Control M

Job description

We are looking for a skilled and motivated Data Platform / DevOps Engineer to operate and evolve our Global Data Platform. In this role, you will work with modern distributed systems and data technologies, ensuring reliable, secure, and scalable data pipelines. You will apply Agile and DevSecOps principles to continuously improve platform stability, automation, and delivery efficiency while collaborating closely with security, engineering, and cloud operations teams.

  • Operate and maintain Global Data Platform components, including:
  • VM servers, Kubernetes clusters, and Kafka
  • Data and analytics applications such as Apache stack, Collibra, Dataiku, and similar tools
  • Design and implement automation for:
  • Infrastructure provisioning
  • Security components
  • CI/CD pipelines supporting ELT/ETL data workflows
  • Build resiliency into data pipelines through:
  • Platform health checks
  • Monitoring and alerting mechanisms
  • Proactive issue prevention and recovery strategies
  • Apply Agile and DevSecOps practices to deliver integrated solutions in iterative increments
  • Collaborate and liaise with:
  • Enterprise Security
  • Digital Engineering
  • Cloud Operations teams
  • Review system issues, incidents, and alerts to:
  • Perform root cause analysis
  • Drive long-term fixes and platform improvements
  • Stay current with industry trends, emerging technologies, and best practices in data platforms and DevOps

Requirements

  • Minimum 5 years of experience designing and supporting large-scale distributed systems
  • Hands-on experience with:
  • Streaming and file-based ingestion (e.g., Kafka, Control-M, AWA)
  • DevOps and CI/CD tooling (e.g., Jenkins, Octopus; Ansible, Chef, XL tools are a plus)
  • Experience with on-premises Big Data architectures; cloud migration experience is an advantage

  • Integration of Data Science workbenches (e.g., Dataiku or similar)

  • Practical experience working in Agile environments (Scrum, SAFe)

  • Supporting enterprise reporting and data science use cases

  • Strong knowledge of modern data architectures:

  • Data lakes, Delta Lakes, Data Meshes, and data platforms
  • Experience with distributed and cloud-native technologies:
  • S3, Parquet, Kafka, Kubernetes, Spark
  • Programming and scripting skills:
  • Python (required)
  • Java / Scala / R
  • Linux scripting, Jinja, Puppet
  • Infrastructure and platform engineering:
  • VM setup and administration
  • Kubernetes scaling and operations
  • Docker, Harbor
  • CI/CD pipelines
  • Firewall rules and security controls

Apply for this position