Skip to content
mimi

Machine Learning Engineer-AI Data Platform

MOBĒ

Reno · On-site Full-time $114k – $130k/yr Today

About the role

Company Overview

MOBE helps people discover new ways to live healthier. We are the whole-person, cross-condition solution that goes further to deliver better health and lower overall costs through evidence-based individual health guidance and pharmacist-led medication management. We empower individuals to make meaningful changes that improve their health and overall well-being. Behind our innovative solutions are robust data analytics, digital application, and a uniquely human philosophy. With one-to-one connection and compassion, we uncover opportunities, overcome challenges, and motivate people to transform their lives.

At MOBE our team is our most significant asset. We cultivate a culture grounded in curiosity, innovation, and growth. We encourage new ideas, fresh solutions, and meaningful impact. We value a workforce made up of people with differences who are eager to learn from each other and grow personally and professionally. We extend this approach to our partners and communities, seeking to increase understanding and expand opportunities across all groups.

Your role at MOBE

We are seeking a highly skilled AI Engineer to serve as a core builder of our AI Data Platform. This role sits at the intersection of machine learning engineering, data platform development, and business intelligence, with responsibility for designing and operating the infrastructure that powers AI-driven insights across the organization.

You will build intelligent data pipelines, production-grade ML systems, and AI-enabled features that translate complex data into actionable outcomes. This role is ideal for an engineer who enjoys working end-to-end from data ingestion and feature engineering to model deployment and downstream consumption in analytics and BI tools. • *Applicants must be authorized to work for ANY employer in the U.S. We are unable to sponsor or take over sponsorship of an employment Visa at this time.

Responsibilities: • Build AI-first data pipelines: Design, implement, and maintain scalable data pipelines that support model training, inference, and analytics use cases across the AI Data Platform. • Deploy production ML systems: Develop, deploy, and monitor machine learning models using AWS SageMaker, ensuring reliability, observability, and performance in production environments. • Implement Retrieval-Augmented Generation (RAG): Architect and maintain RAG-based systems that combine structured and unstructured data to power AI-driven insights and applications. • Operationalize ML lifecycle management: Use MLflow for experiment tracking, model versioning, and lifecycle management to support reproducibility and continuous improvement. • Design feature infrastructure: Build and manage feature stores (e.g., Feast, Tecton, or SageMaker Feature Store) to ensure consistent, reusable features across training and inference. • Orchestrate complex workflows: Create and manage Apache Airflow DAGs to orchestrate data transformations, model pipelines, and AI workflows with clear dependencies and monitoring. • Enable analytics consumption: Partner with BI and analytics teams to ensure ML outputs integrate cleanly with our internal BI reporting hub. • Translate business questions into AI solutions: Collaborate with stakeholders to convert ambiguous business problems into measurable ML- and data-driven solutions. • Uphold data quality and governance: Ensure AI pipelines and models adhere to data governance, security, and quality standards, particularly when handling sensitive data. • Collaborate cross-functionally: Work closely with Data Science, Analytics Engineering, Medical Economics, and DataOps to align AI platform capabilities with business priorities.

Qualifications:

Required: • Five to Seven Years in the ML Engineering space. • Strong proficiency in Python and SQL for data processing, modeling, and pipeline development. • Hands-on experience building and deploying machine learning models in production, including monitoring and performance management. • Experience with AWS-based ML infrastructure, including SageMaker for training, deployment, and inference. • Practical experience designing or operating RAG systems that integrate LLMs with enterprise data sources. • Experience using MLflow (or equivalent) for experiment tracking, model registry, and lifecycle management. • Experience with Apache Airflow for orchestration of data and ML pipelines. • Strong foundation in data engineering concepts, including data modeling, versioning, and testing. • Ability to partner with Med Econ and BI teams to ensure ML outputs are interpretable, trusted, and consumable.

Preferred: • Experience with AWS Bedrock and/or Aider for LLM orchestration or AI-assisted development workflows. • Experience with dbt for transformation modeling, testing, and documentation. • Familiarity with feature store architectures (Feast, Tecton, SageMaker Feature Store). • Experience integrating ML outputs into BI tools such as Tableau, Looker, or QuickSight. • Experience with CI/CD pipelines, Git-based workflows, and infrastructure-as-code practices. • Exposure to healthcare or regulated data environments is a plus but not required.

Nice to Have: • Working knowledge of Docker and Kubernetes for scalable deployment of ML services. • Experience implementing data observability, model drift detection, or AI governance tooling. • Experience fine-tuning or adapting large language models for domain-specific use cases.

Values • People First. We show we care because we believe in the power of human connection • Spark Positivity. We each have the power to turn any challenge into something awesome. • Stay Curious. We relentlessly discover and embrace new ideas to keep moving forward.

Benefits and Compensation

We offer a comprehensive benefits package • Paid time off • Medical, dental, and vision insurance • Life and disability insurance • 401(k) with company match • Tuition reimbursement • Additional benefits available to eligible employees

This is a hybrid position with the expectation of 3 days per week onsite in either our Minneapolis, MN or Reno, NV office.

Requirements

  • This role is ideal for an engineer who enjoys working end-to-end from data ingestion and feature engineering to model deployment and downstream consumption in analytics and BI tools
  • *Applicants must be authorized to work for ANY employer in the U.S. We are unable to sponsor or take over sponsorship of an employment Visa at this time
  • Five to Seven Years in the ML Engineering space
  • Strong proficiency in Python and SQL for data processing, modeling, and pipeline development
  • Hands-on experience building and deploying machine learning models in production, including monitoring and performance management
  • Experience with AWS-based ML infrastructure, including SageMaker for training, deployment, and inference
  • Practical experience designing or operating RAG systems that integrate LLMs with enterprise data sources
  • Experience using MLflow (or equivalent) for experiment tracking, model registry, and lifecycle management
  • Experience with Apache Airflow for orchestration of data and ML pipelines
  • Strong foundation in data engineering concepts, including data modeling, versioning, and testing
  • Ability to partner with Med Econ and BI teams to ensure ML outputs are interpretable, trusted, and consumable
  • Working knowledge of Docker and Kubernetes for scalable deployment of ML services
  • Experience implementing data observability, model drift detection, or AI governance tooling
  • Experience fine-tuning or adapting large language models for domain-specific use cases

Responsibilities

  • This role sits at the intersection of machine learning engineering, data platform development, and business intelligence, with responsibility for designing and operating the infrastructure that powers AI-driven insights across the organization
  • You will build intelligent data pipelines, production-grade ML systems, and AI-enabled features that translate complex data into actionable outcomes
  • Build AI-first data pipelines: Design, implement, and maintain scalable data pipelines that support model training, inference, and analytics use cases across the AI Data Platform
  • Deploy production ML systems: Develop, deploy, and monitor machine learning models using AWS SageMaker, ensuring reliability, observability, and performance in production environments
  • Implement Retrieval-Augmented Generation (RAG): Architect and maintain RAG-based systems that combine structured and unstructured data to power AI-driven insights and applications
  • Operationalize ML lifecycle management: Use MLflow for experiment tracking, model versioning, and lifecycle management to support reproducibility and continuous improvement
  • Design feature infrastructure: Build and manage feature stores (e.g., Feast, Tecton, or SageMaker Feature Store) to ensure consistent, reusable features across training and inference
  • Orchestrate complex workflows: Create and manage Apache Airflow DAGs to orchestrate data transformations, model pipelines, and AI workflows with clear dependencies and monitoring
  • Enable analytics consumption: Partner with BI and analytics teams to ensure ML outputs integrate cleanly with our internal BI reporting hub
  • Translate business questions into AI solutions: Collaborate with stakeholders to convert ambiguous business problems into measurable ML- and data-driven solutions
  • Uphold data quality and governance: Ensure AI pipelines and models adhere to data governance, security, and quality standards, particularly when handling sensitive data
  • Collaborate cross-functionally: Work closely with Data Science, Analytics Engineering, Medical Economics, and DataOps to align AI platform capabilities with business priorities

Benefits

health_insurancepaid_time_offdental_coverage

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free