Lead Software Engineer - Remote
Role details
Job location
Tech stack
Job description
We are seeking a Lead Software Engineer - Remote to design, build, and scale agentic AI systems and advanced ML solutions that improve healthcare delivery and operations. You will architect end-to-end question answering and multiagent workflows, integrate with our internal AI platform (e.g., UAIS), and ensure responsible, compliant use of AI in a regulated environment (HIPAA). The ideal candidate combines deep software engineering expertise with hands on experience in LLMs, RAG, agent/tool use, evaluation on modern cloud and data platforms.
You'll enjoy the flexibility to work remotely * from anywhere within the U.S. as you take on some tough challenges. For all hires in the Minneapolis or Washington, D.C. area, you will be required to work in the office a minimum of four days per week.
Primary Responsibilities:
- Agentic AI Architecture and Delivery
- Design and implement (multi) agentic workflows where LLMs plan, decompose tasks, invoke tools/APIs, and synthesize answers across heterogeneous data sources and services
- Build retrieval augmented generation (RAG) and hybrid search pipelines to power robust question answering over clinical and operational data
- Design, code, test, document, and maintain high quality, scalable Big Data and cloud solutions
- Develop scalable microservices and APIs for integrating agent capabilities into clinician tools and internal apps
- Create prototypes/POCs and conduct design/code reviews to derisk delivery and raise engineering quality
- LLMs, GenAI and Model Adaptation
- Leverage and adapt LLMs; perform prompt engineering, grounding, guard railing, and domain adaptation for healthcare terminology and tasks
- Design intelligent frameworks and finetune models for compliance, accuracy, and ethical standards
- Establish evaluation frameworks (automatic + human in the loop) to measure faithfulness, helpfulness, bias, toxicity, privacy leakage, and overall quality
- Data and Platform Engineering
- Partner with data engineering to build feature/retrieval stores, embeddings pipelines, and ETL/ELT jobs on Spark/Databricks; design analytics models and rules engines
- Define and develop APIs for integrations across the enterprise; improve data access patterns for low latency inference
- Delivery, MLOps and Reliability
- Own MLOps/LLMOps: CI/CD for models/prompts, automated tests (unit/contract/eval), versioning, lineage, rollback; enable blue/green or canary releases
- Instrument SLOs/SLIs (latency, availability, hallucination/defect rate) and cost KPIs (tokens, GPU hours) with dashboards and alerts
- Lead production deployments on internal platforms (e.g., UAIS) with solid observability, reliability, and cost controls
- Security, Privacy and Compliance
- Champion HIPAA and regulated industry controls; integrate access controls, PHI/PPI safeguards, data minimization, encryption, and auditability.
- Collaborate with legal, compliance, and clinical safety to operationalize Responsible AI principles
- Product, Estimation and Collaboration
- Analyze and define customer requirements; assist in defining product technical architecture and delivery roadmaps
- Provide effort estimates and inputs for resource planning; collaborate with QA, architecture, and peer teams
- Write technical documentation, support production, and mentor engineers and data scientists; keep skills current through continuous learning
You'll be rewarded and recognized for your performance in an environment that will challenge you and give you clear direction on what it takes to succeed in your role as well as provide development for other roles you may be interested in.
Requirements
- Bachelor's in Engineering, Computer Science, IT, or related field
- 12+ years of total IT experience
- 8+ years of hands on software development/data engineering/analytics with strong AI/ML delivery (Azure preferred) with Scala, Python, PySpark
- 4+ years of hands on experience with Databricks
- 4+ years of experience with ADF/Airflow (orchestration/scaling)
- 4+ years of experience with bigdata and streaming (Hadoop, MapReduce/HDFS, Spark, Kafka); Docker/Kubernetes
- 4+ years of experience with MySQL and NoSQL databases
- 4+ years of experience with Agile/Scrum, GitHub, Jenkins CI/CD, JUnit; strong coding standards and code reviews
- 2+ years of experience with LLMs and GenAI (Langchain, LangGraph, RAG, Vector DB, Azure Open AI, MCP Server, Agents, LangFuse)
- 2+ years of experience with container (Docker/Kubernetes)
- 1+ years with Proficiency building services or full stack apps (e.g., FastAPI/Flask, Node.js, React/Angular, TypeScript, HTML/CSS), * Healthcare experience; familiarity with clinical datasets
- Experience with SOA and enterprise integration concepts
- Experience working in regulated industries, with knowledge of ethical AI/ML practices and compliance requirements
- Experience with Publications/patents or notable open-source contributions
- Proven excellent analysis, problem solving, and communication skills
All employees working remotely will be required to adhere to UnitedHealth Group's Telecommuter Policy
Benefits & conditions
Pay is based on several factors including but not limited to local labor markets, education, work experience, certifications, etc. In addition to your salary, we offer benefits such as, a comprehensive benefits package, incentive and recognition programs, equity stock purchase and 401k contribution (all benefits are subject to eligibility requirements). No matter where or when you begin a career with us, you'll find a far-reaching choice of benefits and incentives. The salary for this role will range from $112,700 to $193,200 annually based on full-time employment. We comply with all minimum wage laws as applicable.