Senior Data Engineer
Role details
Job location
Tech stack
Job description
We are seeking a talented Senior Data Engineer specializing in Starburst (Trino) and Dell Data Lakehouse to join our AI & Data team. You will be responsible for deploying, maintaining and optimizing Starburst installations & Dell Data Lakehouse, enabling our clients to seamlessly access their data across multiple platforms. The ideal candidate will have excellent communication skills, an advanced understanding of Starburst & Dell Data Lakehouse, and proficiency with troubleshooting and root cause analysis. Responsibilities:
-
Deploy and manage Starburst Enterprise/Galaxy and Dell Data Lakehouse installations, overseeing environment setup, configuration, maintenance, upgrades, and ensuring optimal performance.
-
Configure various server and application settings and parameters.
-
Integrate Starburst with various data sources to create a unified data platform.
-
Design and tune the container solution for performance and scalability.
-
Set up and configure data catalogs in various modes.
-
Implement robust security controls for data access, ensure compliance with data regulations, and manage potential vulnerabilities.
-
Coordinate with various support partners and vendor teams.
-
Troubleshoot and investigate server related issues and provide root cause analysis for incidents.
-
Perform daily server administration and monitoring, and leverage automation (such as Ansible) for efficient maintenance.
-
Plan and execute disaster recovery testing.
-
Create documentation and provide training on Starburst administration and best practices.
Requirements
Do you have experience in System architecture?, Do you have a Bachelor's degree?, * Required Skills & Experience:
- Bachelor's degree in Computer Science, Information Systems, Data Science, Engineering or related field (or equivalent work experience).
- Proven experience with Trino/Starburst Enterprise/Galaxy administration / CLI.
- Implementation experience with container orchestration solutions (Kubernetes/OpenShift).
- Knowledge of Big Data (Hadoop/Hive/Spark) and Cloud technologies (AWS, Azure, GCP).
- Understanding of distributed system architecture, high availability, scalability, and fault tolerance.
- Familiarity with security authentication systems such as LDAP, Active Directory, OAuth2, Kerberos.
- Excellent Unix/Linux skills.
- Familiarity with JDBC / ODBC
- Preferred Skills:
- Certification: Starburst Certified Practitioner.
- Experience Python and/or Java programming.
- Proficient with infrastructure automation tools such as Ansible.
- Knowledge of data requirements for AI and machine learning workloads.
- Familiarity with Data Federation and Cached Services
- Familiarity with Data pipeline (Series of steps that move and transform data from one source to another for analyses and storage)
- Experience with Dell Data Lakehouse administration.
- Experience in Demand Driven Adaptive Enterprise (DDAE) administration
- Working Conditions
- This position may require evening and weekend work for time-sensitive project implementations.