Senior Staff AI Data Engineer - Hybrid

The Hartford
Chicago, United States of America
30 days ago

Role details

Contract type
Permanent contract
Employment type
Full-time (> 32 hours)
Working hours
Regular working hours
Languages
English
Experience level
Senior
Compensation
$ 203K

Job location

Chicago, United States of America

Tech stack

Artificial Intelligence
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Amazon Web Services (AWS)
Data analysis
Computer Vision
Azure
Cloud Computing
Computer Programming
Databases
Continuous Integration
Information Engineering
Data Governance
Data Infrastructure
ETL
Data Systems
Software Design Patterns
DevOps
Disaster Recovery
Amazon DynamoDB
Fault Tolerance
Graph Database
Python
Language Modeling
MongoDB
Natural Language Processing
Neo4j
NoSQL
Open Source Technology
Reliability Engineering
TensorFlow
SQL Databases
Unstructured Data
PyTorch
Large Language Models
Snowflake
Prompt Engineering
Spark
Deep Learning
Generative AI
AWS Lambda
Information Technology
Real Time Data
Data Pipelines

Job description

This role will have a Hybrid work schedule, with the expectation of working in an office location (Hartford, CT; Danbury, CT; Chicago, IL; Columbus, OH; and Charlotte, NC) 3 days a week (Tuesday through Thursday).

Primary Job Responsibilities

  • AI Data Engineering lead responsible for Implementing AI data pipelines that bring together structured, semi-structured and unstructured data to support AI and Agentic solutions. This Includes pre-processing with extraction, chunking, embedding and grounding strategies to get the data ready.

  • Develop AI-driven systems to improve data capabilities, ensuring compliance with industry best practices.

  • Implement efficient Retrieval-Augmented Generation (RAG) architectures and integrate with enterprise data infrastructure.

  • Collaborate with cross-functional teams to integrate solutions into operational processes and systems supporting various functions.

  • Stay up to date with industry advancements in GenAI and apply modern technologies and methodologies to our systems.

  • Design, build and maintain scalable and robust real-time data streaming pipelines using technologies such as SnowPipe.

  • Develop data domains and data products for various consumption archetypes including Reporting, Data Science, AI/ML, Analytics etc

  • Ensure the reliability, availability, and scalability of data pipelines and systems through effective monitoring, alerting, and incident management.

  • Implement best practices in reliability engineering, including redundancy, fault tolerance, and disaster recovery strategies.

  • Collaborate closely with DevOps and infrastructure teams to ensure seamless deployment, operation, and maintenance of data systems.

  • Mentor junior team members and engage in communities of practice to deliver high-quality data and AI solutions while promoting best practices, standards, and adoption of reusable patterns.

  • Develop graph database solution for complex data relationships supporting AI systems.

  • Apply GenAI solutions to insurance-specific data use cases and challenges.

  • Partner with architects and stakeholders to influence and implement the vision of the AI and data pipelines while safeguarding the integrity and scalability of the environment.

Requirements

  • Strong Technical Knowledge (AI solution leveraging Cloud and modern solutions)

  • Able to communicate effectively with both technical and non-technical teams and influence

  • Collaboration across teams, decision making, conflict resolution and relationship building skills.

  • Experience in mentoring and developing Junior AI or Data Engineers.

  • Knowledge of evolving industry design patterns for AI.

  • Strong planning, organization, and execution skills.

  • Ability to provide Thought Leadership to dynamic and collaborative teams, demonstrating excellent interpersonal skills and time management capabilities.

  • Ability to understand and align deliverables to the departmental and organization strategies and objectives.

  • Ability to lead successfully in a lean, agile, and fast-paced organization, leveraging Scaled Agile principles and ways of working.

  • Leader and team player with a transformation mindset.

  • Ability to translate complex technical topics into business solutions and strategies, as well as turn business requirements into a technical solution.

Qualifications

  • Candidates must be authorized to work in the US without company sponsorship. The company will not support the STEM OPT I-983 Training Plan endorsement for this position.

  • Bachelor's in Computer Science, Artificial Intelligence, or a related field.

  • 8+ years of data engineering experience including Data solutions, SQL and NoSQL, Snowflake, ETL/ELT tools, CICD, Python/Spark, Datamesh or Data Fabric.

  • 1+ years of data engineering experience focused on supporting Generative AI technologies

  • Strong hands-on experience implementing production ready enterprise grade GenAI data solutions.

  • Experience with prompt engineering techniques for large language models.

  • Experience in implementing Retrieval-Augmented Generation (RAG) pipelines, integrating retrieval mechanisms with language models.

  • Experience of vector databases and graph databases, including implementation and optimization.

  • Experience in processing and leveraging unstructured data for GenAI applications.

  • Proficiency in implementing scalable AI driven data systems supporting agentic solution (AWS Lambda, S3, EC2, Lang chain, Langgraph).

  • Strong programming skills in Python and familiarity with deep learning frameworks such as PyTorch or TensorFlow.

  • Experience with building AI pipelines that bring together structured, semi-structured and unstructured data. This includes pre-processing with extraction, chunking, embedding and grounding strategies, semantic modeling, and getting the data ready for Models and Agentic solutions.

  • Experience in vector databases, graph databases, NoSQL, Document DBs, including design, implementation, and optimization. (e.g., AWS open search, GCP Vertex AI, Neo4j, Spanner Graph, Neptune, Mongo, DynamoDB etc.).

  • Experience in implementing data governance practices, including Data Quality, Lineage, Data Catalogue capture, holistically, strategically, and dynamically on a large-scale data platform.

  • Experience with cloud platforms (AWS or GCP or Azure)

  • Strong written and verbal communication skills and ability to explain technical concepts to various stakeholders.

Preferred Qualifications:

  • Experience in multi cloud hybrid AI solutions.

  • AI Certifications

  • Experience in P&C industry

  • Knowledge of natural language processing (NLP) and computer vision technologies.

  • Contributions to open-source AI projects or research publications in the field of Generative AI.

Benefits & conditions

The listed annualized base pay range is primarily based on analysis of similar positions in the external market. Actual base pay could vary and may be above or below the listed range based on factors including but not limited to performance, proficiency and demonstration of competencies required for the role. The base pay is just one component of The Hartford's total compensation package for employees. Other rewards may include short-term or annual bonuses, long-term incentives, and on-the-spot recognition. The annualized base pay range for this role is:

$135,040 - $202,560

About the company

We're determined to make a difference and are proud to be an insurance company that goes well beyond coverages and policies. Working here means having every opportunity to achieve your goals - and to help others accomplish theirs, too. Join our team as we help shape the future.

Apply for this position