ETL Developer

Opus Recruitment Solutions
Charing Cross, United Kingdom
2 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Junior
Compensation
£ 55K

Job location

Remote
Charing Cross, United Kingdom

Tech stack

Airflow
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Big Data
Cloud Storage
Data as a Services
Information Engineering
ETL
Data Manipulation Languages
Data Migration
Data Mining
Data Warehousing
Distributed Systems
MapReduce
Python
NumPy
Oracle Business Intelligence Enterprise Edition
Azure
SQL Stored Procedures
SQL Databases
Technical Data Management Systems
Data Processing
Scripting (Bash/Python/Go/Ruby)
Azure
Snowflake
Spark
GIT
Pandas
Data Lake
PySpark
Information Technology
Amazon Web Services (AWS)
Star Schema
Google BigQuery
Software Version Control
Data Pipelines

Job description

A leading tech consultancy is seeking a Data Engineer to support a university's major data platform transformation project using AWS. The successful candidate will have strong hands-on experience with AWS Glue and PySpark, responsible for developing scalable data workflows...

Requirements

!! IMMEDIATE JOINERS !! Junior Big Data Developer (Python & SQL Focus) We're looking for an enthusiastic and detail-oriented Junior Big Data Developer to join our data engineering team. This role is ideal for an early-career professional with foundational knowledge in data processing, strong proficiency in Python, and expert skills in SQL. You'll focus on building, testing, and maintaining data pipelines and ensuring data quality across our scalable Big Data platforms.Key ResponsibilitiesData Pipeline Development: Assist in the design, construction, and maintenance of robust ETL/ELT pipelines to integrate data from various sources into our data warehouse or data lake.Data Transformation with Python: Write, optimize, and maintain production-grade Python scripts to clean, transform, aggregate, and process large volumes of data.Database Interaction (SQL): Develop complex, high-performance SQL queries (DDL/DML) for data extraction, manipulation, and validation within relational and data warehousing environments.Quality Assurance: Implement data quality checks and monitoring across pipelines, identifying discrepancies and ensuring the accuracy and reliability of data.Collaboration: Work closely with Data Scientists, Data Analysts, and other Engineers to understand data requirements and translate business needs into technical data solutions.Tooling & Automation: Utilize version control tools like Git and contribute to the automation of data workflows and recurring processes.Documentation: Create and maintain technical documentation for data mappings, processes, and pipelines.Required Skills and QualificationsCore Technical SkillsSkill AreaRequirements ProgrammingStrong proficiency in Python for data manipulation and scripting. Familiarity with standard Python data libraries (e.G., Pandas, NumPy).DatabaseExpert-level proficiency in SQL (Structured Query Language). Experience writing complex joins, stored procedures, and performing performance tuning.Big Data ConceptsFoundational understanding of Big Data architecture (Data Lakes, Data Warehouses) and distributed processing concepts (e.G., MapReduce).ETL/ELTBasic knowledge of ETL principles and data modeling (star schema, snowflake schema).Version ControlPractical experience with Git (branching, merging, pull requests).Preferred Qualifications (A Plus)Experience with a distributed computing framework like Apache Spark (using PySpark).Familiarity with cloud data services (AWS S3/Redshift, Azure Data Lake/Synapse, or Google BigQuery/Cloud Storage).Exposure to workflow orchestration tools (Apache Airflow, Prefect, or Dagster).Bachelor's degree in Computer Science, Engineering, Information Technology, or a related field. Similar jobs, A private equity investor in London seeks a Data Engineer to enhance its data platform. You'll own key components, develop ETL pipelines, and lead data migration. Applicants should have 4-6 years of data engineering experience and proficiency in Python and SQL, ideally with...

About the company

A leading data solutions company in the United Kingdom is seeking a Data Architect skilled in designing scalable data solutions using the Snowflake platform. This role involves developing ETL/ELT pipelines, collaborating with teams to meet data requirements, and..., Job Description Jameson Legal is working with a highly regarded UK law firm seeking a Senior Associate to join its Projects, Development and Construction team in London. This is an exceptional opportunity for an experienced construction lawyer to work on major...

Apply for this position