Data Engineer

Globant
4 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English, German

Job location

Remote

Tech stack

Java
Airflow
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Big Data
Cloudera Impala
Computer Programming
Databases
Data Integration
ETL
Data Transformation
Data Systems
Software Debugging
Hadoop
Hive
Python
Cloudera
Scala
SQL Databases
Data Processing
Spark
Amazon Web Services (AWS)
Apache Flink
Kafka
Apache Nifi
Splunk
New Relic (SaaS)
Data Pipelines

Job description

  • Design and implement end-to-end data pipelines, ensuring seamless data transformation and transfer across systems.
  • Build cloud-based data solutions on AWS using Kinesis, Flink, Glue, ECS, S3, and VPC Endpoint.
  • Develop robust, efficient data processing applications in Java, Python, and Scala.
  • Create, manage, and optimize workflows for large-scale data processing with Apache Spark, NiFi, Airflow, and SQL.
  • Efficiently organize and access large amounts of data using Iceberg, Impala, Hive, Hadoop, and Kafka.
  • Ensure the reliability of data pipelines and system performance with monitoring tools such as OpenTelemetry, New Relic, and Splunk.
  • Establish ETL processes, database architecture, and management for efficient data integration and organization.
  • Implement scalable Big Data solutions with Apache Spark, Hadoop, and Cloudera.
  • Analyze, debug, and resolve complex data-related challenges.
  • Work effectively in cross-functional teams and clearly communicate technical solutions to various stakeholders.

Requirements

Do you have experience in Splunk?, * Experience with AWS services: Kinesis, Flink, Glue, ECS, S3, and VPC Endpoint for scalable, reliable cloud-based data solutions.

  • Fluent in German and English.
  • Strong programming proficiency in Java, Python, and Scala for robust, efficient data processing.
  • Expertise in Apache Spark, NiFi, Airflow, and SQL for creating, managing, and optimizing large-scale workflows.
  • Knowledge of storage systems: Iceberg, Impala, Hive, Hadoop, and Kafka.
  • Familiarity with monitoring tools such as OpenTelemetry, New Relic, and Splunk to ensure pipeline reliability and system performance.
  • Comprehensive understanding of ETL processes, database architecture, and management.
  • Proven ability to design and implement end-to-end data pipelines.
  • Hands-on experience with Apache Spark, Hadoop, and Cloudera for scalable Big Data solutions.
  • Excellent analytical and troubleshooting skills.
  • Effective collaboration in cross-functional teams and clear communication of technical content to stakeholders.

Benefits & conditions

Final compensation offered is based on multiple factors such as the specific role, hiring location, as well as individual skills, experience, and qualifications. In addition to competitive salaries, we offer a comprehensive benefits package. L earn more about life at Globant here: Globant Experience Guide .

About the company

At Globant, we are working to make the world a better place, one step at a time. We enhance business development and enterprise solutions to prepare them for a digital future. With a diverse and talented team present in more than 30 countries, we are strategic partners to leading global companies in their business process transformation.

Apply for this position