<Back to Search
Principal AI or Machine Learning Data Engineer - Remote or hybrid from MN or DC
Eden Prairie, MNMarch 26th, 2026
Optum is a global organization that delivers care, aided by technology to help millions of people live healthier lives. The work you do with our team will directly improve health outcomes by connecting people with the care, pharmacy benefits, data and resources they need to feel their best. Here, you will find a culture guided by diversity and inclusion, talented peers, comprehensive benefits and career development opportunities. Come make an impact on the communities we serve as you help us advance health equity on a global scale. Join us to start Caring. Connecting. Growing together.The Enterprise Information Security (EIS) team is responsible for cybersecurity across our organization. We support our business and members by reducing risk, rapidly responding to threats, focusing on business resiliency and securing new acquisitions.The Principal AI / Machine Learning Data Engineer role focuses on designing and building scalable data platforms that enable advanced analytics, machine learning, and AI-driven solutions. This role will support the development of intelligent systems that process large-scale event and operational data, enabling faster insights, automation, and decision-making across the organization.This position sits at the intersection of data engineering, machine learning, and AI, with an emphasis on building modern data pipelines and enabling production-grade AI capabilities.Ideal Candidate Profile:Demonstrated experience building and operating production data platforms and pipelines across batch and streaming workloadsSolid hands-on engineering in Python and SQL; familiarity with JVM languages (Java/Scala) in Spark ecosystems is a plusExperience with distributed processing and lakehouse/warehouse patterns (eg, Spark/PySpark, Databricks, Snowflake)Experience building ingestion frameworks for structured and unstructured data, including event/log and semi-structured formatsExperience enabling Generative AI solutions in production (eg, RAG-style architectures), including retrieval patterns and evaluation/monitoring practicesFamiliarity with knowledge-centric data approaches (eg, metadata-driven systems, entity resolution, and/or graph concepts) to improve discoverability and downstream analyticsSolid data quality, observability, and monitoring mindset (profiling, validation, alerting, and reliability improvements)Comfort with orchestration, CI/CD, containerization, and infrastructure-as-code (eg, Airflow, GitHub Actions, Docker, Terraform, Kubernetes)Cloud experience (AWS, Azure, and/or GCP), including secure handling of sensitive data (PII/PHI) and collaboration with compliance partnersAbility to lead through influence, mentor engineers, and translate ambiguous problems into scalable technical roadmapsYou'll enjoy the flexibility to work remotely * from anywhere within the U.S. as you take on some tough challenges. For all hires in the Minneapolis or Washington, D.C. area, you will be required to work in the office a minimum of four days per week.Primary Responsibilities:Design, develop, and maintain scalable data pipelines and data platforms supporting analytics, machine learning, and AI use casesBuild and optimize ingestion frameworks for large-scale structured and unstructured data, including streaming and event-driven sourcesPartner with cross-functional stakeholders to understand evolving data and AI needs and define long-term technical solutionsEnable and support machine learning and AI workflows, including feature engineering, data preparation, and model deployment supportDrive strategic initiatives around Generative AI, data quality, observability, lineage, and governanceDevelop and maintain frameworks that support rapid experimentation and deployment of AI/ML solutionsIntroduce and evolve best practices in data modeling, orchestration, testing, and monitoringIdentify and champion opportunities for platform scalability, performance optimization, and cost efficiencyCollaborate with product, analytics, and infrastructure teams to deliver high-impact data and AI solutionsBuild and maintain reusable parsing, enrichment, analytic, and service libraries to accelerate delivery across teamsWork comfortably under time-sensitive conditions while ensuring thoroughnessMaintain high ethical standards and the ability to remain objective and confidentialYou'll be rewarded and recognized for your performance in an environment that will challenge you and give you clear direction on what it takes to succeed in your role as well as provide development for other roles you may be interested in.Required Qualifications:Bachelor's degree or equivalent experience5+ years of experience designing, building, and operating production data pipelines and platforms5+ years of hands-on development with Python (preferred) and/or Java, including code reviews, packaging, and deployment5+ years of experience with Spark (PySpark) and Databricks (or similar distributed data processing platform)2+ years of experience leveraging and deploying Generative AI use cases to production environmentsSolid SQL skills and experience working with data lakes and warehouses (e.g., Databricks, Snowflake)Experience building ingestion frameworks for structured and unstructured data (e.g., event/log, semi-structured JSON), including parsing and enrichment patternsExperience designing and scaling ELT/ETL frameworks with orchestration tools such as Airflow (or equivalent)Experience implementing data quality, observability, and monitoring practices (e.g., data quality checks, pipeline SLAs/SLOs, alerting)Experience with metadata, lineage, and governance concepts and tooling (e.g., data catalogs, lineage, access controls)Experience with data modeling best practices for analytics and ML use casesExperience with DevOps and CI/CD practices and tools (e.g., GitHub Actions), containerization, and infrastructure-as-code (e.g., Docker, Kubernetes, Terraform)Experience supporting ML/AI workflows (feature engineering, data preparation, and model deployment enablement); exposure to MLOps practices is a plusDemonstrated ability to partner with cross-functional stakeholders, translate requirements into technical solutions, and lead through influencePreferred Qualifications:Experience with cloud platforms such as AWS, Azure, or Google Cloud, including managed data servicesExperience with streaming and event-driven architectures (e.g., Kafka, Kinesis, Event Hubs)Experience with data quality and validation frameworks (e.g., Great Expectations, Deequ) and/or data observability toolingExperience enabling MLOps practices (e.g., feature stores, model registries, experiment tracking, deployment automation)Experience with lakehouse architectures, Delta Lake, and advanced Spark optimization/performance tuningExperience with data visualization tools and libraries such as Plotly, seaborn, and ChartjsExperience with machine learning and predictive analyticsFamiliarity with security and privacy concepts for data platforms (e.g., least privilege, PII/PHI handling) and working with compliance partners*All employees working remotely will be required to adhere to UnitedHealth Group's Telecommuter PolicyPay is based on several factors including but not limited to local labor markets, education, work experience, certifications, etc. In addition to your salary, we offer benefits such as, a comprehensive benefits package, incentive and recognition programs, equity stock purchase and 401k contribution (all benefits are subject to eligibility requirements). No matter where or when you begin a career with us, you'll find a far-reaching choice of benefits and incentives. The salary for this role will range from $112,700 to $193,200 annually based on full-time employment. We comply with all minimum wage laws as applicable.Application Deadline: This will be posted for a minimum of 2 business days or until a sufficient candidate pool has been collected. Job posting may come down early due to volume of applicants.At UnitedHealth Group, our mission is to help people live healthier lives and make the health system work better for everyone. We believe everyone-of every race, gender, sexuality, age, location and income-deserves the opportunity to live their healthiest life. Today, however, there are still far too many barriers to good health which are disproportionately experienced by people of color, historically marginalized groups and those with lower incomes. We are committed to mitigating our impact on the environment and enabling and delivering equitable care that addresses health disparities and improves health outcomes - an enterprise priority reflected in our mission. UnitedHealth Group is an Equal Employment Opportunity employer under applicable law and qualified applicants will receive consideration for employment without regard to race, national origin, religion, age, color, sex, sexual orientation, gender identity, disability, or protected veteran status, or any other characteristic protected by local, state, or federal laws, rules, or regulations.UnitedHealth Group is a drug - free workplace. Candidates are required to pass a drug test before beginning employment.
Showing 200 of 18,123 matching similar jobs in Springbrook, ND
- Lead Data Engineer
- Senior ETL Tester: Data Quality & Automation Lead
- Senior Data Engineer (Snowflake) - Onsite
- Senior Data Engineer: Cloud ETL & Analytics
- Data Engineer (Mid-level)
- Senior Data Analytics Engineer-Databricks & Power BI Austin
- Data Engineer: Build Scalable Healthcare Analytics
- Data Engineer: ETL Pipelines & Azure Data Factory Pro
- Data Engineer
- Lead Data Engineer- Ataccama
- Senior Data Engineer - Scalable Pipelines & Analytics
- Lead Data Engineer: Build Scalable Modern Data Platform
- Senior GCP Data Engineer - PySpark & BigQuery Expert
- Senior Data Engineer - Java Spring Batch ETL (Hybrid)
- Associate Principal Engineer, Big Data
- Snowflake Data Engineer
- Analytics Developer I - Business Intelligence
- Senior Data Engineer - Real-time ETL & Data Platform Lead
- Slalom Flex (Project Based)- Federal GCP Data Engineer
- Slalom Flex (Project Based)- Federal GCP Data Engineer
- Sr. Analytics Engineer
- Slalom Flex (Project Based)- Federal GCP Data Engineer
- Slalom Flex (Project Based)- Federal GCP Data Engineer
- Slalom Flex (Project Based)- Federal GCP Data Engineer
- Sr. BI/DW Engineer
- Advisory Lead, AI/Data Due Diligence
- Sr. BI/DW Engineer
- Sr. BI/DW Engineer
- Sr. BI/DW Engineer
- Sr. BI/DW Engineer
- Sr. BI/DW Engineer
- Sr. BI/DW Engineer
- Lead Data Scientist
- Data Architect - Power & Utilities - Senior Manager- Consulting - Location OPEN
- Data Architect
- Data Architect
- Junior java microservices developer/Jr data analyst
- Agentic AI Engineer
- Data Engineer
- Worldwide Specialist Solutions Architect - GenAI, Data & AI GTM