Data Engineer

JD Sports
Bury, United Kingdom
2 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Junior
Compensation
£ 44K

Job location

Bury, United Kingdom

Tech stack

Artificial Intelligence
Data analysis
Azure
Big Data
Data Normalization
Python
SQL Databases
Data Streaming
GIT
PySpark

Job description

Working with a wide array of tools, you will work through the full, streaming, enrichment & curation of data into a cloud-based data environment, verifying data, and making sure that links to other key data sets are obtained to allow for simple, effective data analysis for our insight team and data scientists. The role will allow you to start looking at whether you would like to become a specialist in a specific field such as infrastructure, curation or ML/AI in order to further progress you career. Responsibilities: Assisting in the automation & maintenance of pipelines within a cloud-based environment, the candidate would also be involved in the sourcing of data using a range of different methods, whilst carrying out verification that the data is acceptable for ingestion.

Analysis of large data sets using tools such as Python & SQL Setting up new pipelines for the full stream/enrichment/curation process Upkeep of source code locations

Role objectives and KPI's:

Analysis of large data sets using tools such as Python & SQL Creation stream/enrichment/curation process utilising a wide variety of data sources Upkeep of source code locations/Git Hub Repositories Set up of tables/views/procedures Data aggregation & manipulation Building of large scale analytical data sets Investigation of new/alternative technology

Competencies and behaviours:

Deciding and Initiating Action 2.1 Working with People 3.1 Relating and Networking 4.2 Applying Expertise and Technology 5.1 Learning and Researching 6.1 Planning and Organising 8.1 Achieving Personal Work Goals and Objectives

Requirements

An excellent level of experience in tools such as SQL / Python Ability to decide on the overall concept and vision Creating road maps 1-2 years of cloud data solution experience in GCP / Azure Experience of Pyspark coding or equivalent Excellent problem solving skills Strong attention to detail Strong stakeholder management Strong communication skills

Apply for this position