Senior Data Engineer
Role details
Job location
Tech stack
Job description
We are seeking a talented Senior Data Engineer specializing in Starburst (Trino) and Dell Data Lakehouse to join our AI & Data team. You will be responsible for deploying, maintaining and optimizing Starburst installations & Dell Data Lakehouse, enabling our clients to seamlessly access their data across multiple platforms. The ideal candidate will have excellent communication skills, an advanced understanding of Starburst & Dell Data Lakehouse, and proficiency with troubleshooting and root cause analysis., * Deploy and manage Starburst Enterprise/Galaxy and Dell Data Lakehouse installations, overseeing environment setup, configuration, maintenance, upgrades, and ensuring optimal performance.
- Configure various server and application settings and parameters.
- Integrate Starburst with various data sources to create a unified data platform.
- Design and tune the container solution for performance and scalability.
- Set up and configure data catalogs in various modes.
- Implement robust security controls for data access, ensure compliance with data regulations, and manage potential vulnerabilities.
- Coordinate with various support partners and vendor teams.
- Troubleshoot and investigate server related issues and provide root cause analysis for incidents.
- Perform daily server administration and monitoring, and leverage automation (such as Ansible) for efficient maintenance.
- Plan and execute disaster recovery testing.
- Create documentation and provide training on Starburst administration and best practices.
Requirements
Do you have a strong background in Data Engineering with a high level understanding of Trino/Starburst Enterprise, along with Big Data, Cloud tech, Kubernetes/OpenShift and Unix/Linux skills?, * Bachelor's degree in Computer Science, Information Systems, Data Science, Engineering or related field (or equivalent work experience).
- Proven experience with Trino/Starburst Enterprise/Galaxy administration/CLI.
- Implementation experience with container orchestration solutions (Kubernetes/OpenShift).
- Knowledge of Big Data (Hadoop/Hive/Spark) and Cloud technologies (AWS, Azure, GCP).
- Understanding of distributed system architecture, high availability, scalability, and fault tolerance.
- Familiarity with security authentication systems such as LDAP, Active Directory, OAuth2, Kerberos.
- Excellent Unix/Linux skills.
- Familiarity with JDBC/ODBC
Preferred Skills:
- Certification: Starburst Certified Practitioner.
- Experience Python and/or Java programming.
- Proficient with infrastructure automation tools such as Ansible.
- Knowledge of data requirements for AI and machine learning workloads.
- Familiarity with Data Federation and Cached Services
- Familiarity with Data pipeline (Series of steps that move and transform data from one source to another for analyses and storage)
- Experience with Dell Data Lakehouse administration.
- Experience in Demand Driven Adaptive Enterprise (DDAE) administration
Benefits & conditions
- This position may require evening and weekend work for time-sensitive project implementations.
Some of the benefits include:
- Healthcare and dental insurance
- Company pension is matched up to 5%
- 25 days annual leave entitlement plus bank holidays and the option to purchase 5 extra days
- Life assurance - 4 x annual salary
- Cycle to work scheme
- Client prioritises internal development opportunities and offer access to our Udemy training platform with over 5000 training courses
Disclaimer