AI or ML Engineer - Remote
Mentions vibe coding tools explicitly (Cursor, Claude Code, Replit) in preferred qualifications.
About the Role
Optum is hiring an AI/ML Engineer to design, build, and operate scalable machine learning and generative AI platforms for production use across the enterprise. The role focuses on MLOps/LLMOps, model lifecycle management, cloud-native infrastructure, and collaboration with data science and engineering teams; it is primarily remote within the U.S. with local office requirements for Minneapolis and Washington, D.C. hires.
Job Description
Role
Optum is seeking an AI/ML Engineer to build and operate enterprise-scale machine learning and generative AI platforms. The role centers on designing and implementing ML infrastructure, model lifecycle management, MLOps/LLMOps pipelines, and productionizing models for batch and real-time inference across the organization.
Key Responsibilities
- Build enterprise ML and GenAI platforms to support experimentation, training, evaluation, deployment, monitoring, and lifecycle management.
- Productionize ML and generative AI models using batch and real-time inference architectures.
- Develop and operate MLOps and LLMOps pipelines, including CI/CT/CD workflows for model testing, validation, versioning, and promotion.
- Create scalable, cloud-native ML infrastructure using containers and orchestration platforms.
- Build and manage LLM application stacks (gateways, orchestration, model routing, caching) and optimize cost/performance.
- Implement monitoring and lifecycle systems to track drift, latency, bias, and data quality and enable automated retraining.
- Ensure governance, security, auditability, reproducibility, and observability of ML systems.
- Partner cross-functionally with data scientists, data engineers, and software engineers to define production standards and scalable AI solutions.
Requirements
- 5+ years experience in machine learning engineering, MLOps, or AI platform engineering building production ML systems and scalable pipelines.
- 4+ years programming in Python for ML systems; familiarity with frameworks such as PyTorch, TensorFlow, or scikit-learn.
- 3+ years with ML lifecycle platforms (e.g., MLflow, Kubeflow, SageMaker, Azure ML, GCP Vertex AI).
- 3+ years building cloud-native ML platforms using Docker, Kubernetes, and distributed systems.
- 3+ years with distributed data processing and orchestration tools such as Spark, Ray, Airflow, Dagster, or Prefect.
- 3+ years working with Generative AI and LLMs, including prompt engineering, prompt chaining, and fine-tuning (instruction tuning and LoRA/qLoRA).
Preferred Qualifications
- Master’s degree in Computer Science, Engineering, Data Science, or related discipline.
- Experience with LLM orchestration frameworks (e.g., LangChain, LlamaIndex, Semantic Kernel).
- Experience in regulated or enterprise environments with emphasis on security, compliance, and responsible AI.
- Experience with vibe coding tools (Cursor, Claude Code, Replit).
- Experience operating multi-cloud or hybrid ML platforms, LLM cost optimization, knowledge graphs or hybrid search, and contributions to open-source ML/MLOps tooling.
Location & Remote
- Remote work is available from anywhere within the U.S. Hires based in Minneapolis or Washington, D.C. are required to work in-office a minimum of four days per week.
- Remote hires must adhere to UnitedHealth Group’s Telecommuter Policy.
Compensation & Benefits
- Salary range: $98,547 to $175,978 annually (based on full-time employment).
- Benefits noted include a comprehensive benefits package, incentive and recognition programs, equity stock purchase, and 401(k) contribution (subject to eligibility).
Additional Notes
- Role requires passing a pre-employment drug test and compliance with applicable company policies.
- Focus is on scalable, secure, and governed ML systems for healthcare use cases.