Data Engineer

Infinity Tech Group Inc
Princeton, United States of America
yesterday

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Intermediate

Job location

Princeton, United States of America

Tech stack

API
Airflow
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Azure
Big Data
Google BigQuery
Cloud Computing
Computer Programming
Databases
Information Engineering
Data Governance
ETL
Data Mart
Data Security
Data Structures
Data Systems
Data Warehousing
Database Queries
Python
PostgreSQL
MongoDB
MySQL
NoSQL
NumPy
Data Streaming
Data Processing
Google Cloud Platform
Azure
Snowflake
GIT
Pandas
Data Lake
PySpark
Kafka
Spark Streaming
Data Management
Machine Learning Operations
Data Pipelines
Docker
Databricks

Job description

We are looking for a skilled Data Engineer with strong expertise in Python, data processing, and modern cloud data platforms like Snowflake or Databricks. The role involves building scalable data pipelines, enabling analytics, and supporting data-driven decision-making across the organization., * Design, build, and maintain scalable data pipelines (ETL/ELT)

  • Develop data workflows using Python
  • Work with large datasets in structured and unstructured formats
  • Implement data solutions using Snowflake or Databricks
  • Optimize data pipelines for performance, reliability, and cost
  • Build and manage data models, data marts, and data lakes
  • Integrate data from multiple sources (APIs, databases, streaming)
  • Ensure data quality, governance, and security
  • Collaborate with data analysts, scientists, and business teams
  • Automate workflows using orchestration tools (Airflow, etc.)

Requirements

Programming & Data

  • Strong proficiency in Python
  • Experience with:
  • Pandas, NumPy, PySpark
  • Solid understanding of:
  • Data structures & algorithms
  • Data processing techniques

Data Engineering

  • Hands-on with:
  • ETL/ELT pipeline development
  • Data warehousing concepts
  • Batch & streaming data processing

Platforms

  • Experience with at least one:
  • Snowflake, * Strong SQL skills
  • Experience with:
  • Relational DBs (PostgreSQL, MySQL)
  • NoSQL (MongoDB - optional)

Cloud (Preferred)

  • Experience in:
  • AWS (S3, Glue, Redshift)
  • Azure (ADF, ADLS)
  • Google Cloud Platform (BigQuery), * Apache Airflow / Prefect
  • Git, CI/CD pipelines
  • Docker (good to have), * 3-8+ years in Data Engineering or similar role, * Snowflake SnowPro Certification
  • Databricks Certified Data Engineer
  • AWS / Azure Data certifications, * Experience with real-time streaming (Kafka, Spark Streaming)
  • Knowledge of data governance & catalog tools
  • Exposure to Machine Learning pipelines
  • Understanding of data security & compliance, * Strong analytical and problem-solving mindset
  • Ability to work with cross-functional teams
  • Clear communication and documentation skills

Apply for this position