Skip to content
mimi

Data Engineer – Data Access & Delivery

OneMedNet

Remote (Global) Full-time Mid Level CA$80k – CA$120k/yr Today

About the role

Location

Vancouver, BC (Remote) / Canada (Remote)

Salary

$80K – $120K

About

OneMedNet is the leading provider of Real World Data (RWD) and imaging Real World Data (iRWD), supplying highly curated, regulatory‑grade clinical and medical imaging data to advance life sciences. Continually expanding our extensive provider network is an integral part of OneMedNet’s strategy and a cornerstone to delivering value. We leverage cutting‑edge technologies to unlock RWD for life sciences, pharmacy, and device research and development organizations. Committed to privacy and compliance, we exceed all relevant HIPAA, and other privacy laws to ensure the security and confidentiality of patient information.

As a member of the Data Access & Delivery team, you will build the end‑to‑end engine responsible for the discovery, retrieval, and delivery of clinical data. We are looking for a high‑impact engineer with expert‑level SQL and Python skills and a proven ability to automate manual workflows within Cloud environments. You will play a critical role in engineering scalable, de‑identified data pipelines that maintain the highest standards of clinical integrity and security.

The ideal candidate is comfortable with healthcare concepts and eager to develop deep expertise in clinical terminologies (ICD‑10, CPT, SNOMED, LOINC, RxNorm) and large‑scale healthcare data models, including OMOP, HL7 FHIR, EPIC, Cerner, and DICOM. You will work across these standards to extract, transform, and deliver clinical data that powers life sciences research.

Duties & Responsibilities

  • Data Extraction & Integration: Build scalable pipelines to extract diverse healthcare data, including structured and non‑structured clinical records, medical imaging, and standardized health messaging.
  • Technical Discovery: Execute complex queries to validate data availability and volume across partner systems to support opportunities.
  • Automated De‑ID: Architect and maintain automated de‑identification protocols using state‑of‑the‑art tools, including symbolic and statistical AI, LLMs, and SLMs.
  • Orchestration & Staging: Leverage the AWS ecosystem and modern data platforms to orchestrate secure data migrations and manage multi‑stage staging environments.
  • Workflow Automation: Design and implement cross‑platform tools to automate manual curation steps, significantly increasing data velocity.
  • Systems Reliability: Maintain high‑performance pipeline stability and redundancy.
  • Cross‑Functional Collaboration: Work closely with Sales, Product, and Engineering teams to gather requirements, design solutions, and deliver high‑quality data products.
  • Continuous Improvement: Stay current on emerging data technologies and contribute to the evolution of development processes and best practices.
  • Other duties as assigned.

Qualifications

  • Bachelor’s or Master’s degree in Computer Science, Data Science, Health Informatics, or a related field.
  • Minimum 2 years of professional experience in data engineering, with strong proficiency in SQL and Python for healthcare data engineering and task automation.
  • Familiarity with healthcare data standards and willingness to develop deep expertise in clinical terminologies (ICD‑10, CPT, SNOMED, LOINC, RxNorm) and data models, including OMOP, HL7 FHIR, EPIC, Cerner, and DICOM.
  • Comfortable with healthcare concepts and clinical terminology; experience in the medical data or healthcare industry is a plus.
  • Experience with LLMs or AI/ML pipelines is a plus.
  • Understanding of technical de‑identification techniques (PHI/PII removal) for sensitive clinical information.
  • Hands‑on experience with the AWS ecosystem (S3, EC2, RDS, VPC, IAM) and modern data/AI platforms (e.g., Snowflake, Palantir).
  • Proficiency in Linux CLI and practical experience building, deploying, and managing Docker containers.
  • Proven ability to identify manual bottlenecks and replace them with scalable, engineered solutions.
  • Excellent problem‑solving skills and attention to detail, with a focus on delivering high‑quality data solutions.
  • Effective communication skills and ability to collaborate with cross‑functional teams.
  • Ability to work independently and as part of a team in a fast‑paced, dynamic environment.

Pay & Benefits

  • Competitive salary
  • Excellent Medical, Dental, & Vision Insurance
  • Flexible work schedule
  • Unlimited PTO
  • Remote office flexibility
  • Fun team at an exciting time in an emerging market

Join us in our mission to revolutionize healthcare through innovative data solutions. If you are passionate about making a meaningful impact and possess the skills and experience outlined above, we’d love to hear from you. Apply now and be a part of our dynamic team!

#J-18808-Ljbffr

Requirements

  • Minimum 2 years of professional experience in data engineering, with strong proficiency in SQL and Python for healthcare data engineering and task automation.
  • Understanding of technical de-identification techniques (PHI/PII removal) for sensitive clinical information.
  • Hands-on experience with the AWS ecosystem (S3, EC2, RDS, VPC, IAM) and modern data/AI platforms (e.g., Snowflake, Palantir).
  • Proficiency in Linux CLI and practical experience building, deploying, and managing Docker containers.
  • Proven ability to identify manual bottlenecks and replace them with scalable, engineered solutions.
  • Excellent problem-solving skills and attention to detail, with a focus on delivering high-quality data solutions.
  • Effective communication skills and ability to collaborate with cross-functional teams.
  • Ability to work independently and as part of a team in a fast-paced, dynamic environment.

Responsibilities

  • Build scalable pipelines to extract diverse healthcare data, including structured and non-structured clinical records, medical imaging, and standardized health messaging.
  • Execute complex queries to validate data availability and volume across partner systems to support opportunities.
  • Architect and maintain automated de-identification protocols using state-of-the-art tools, including symbolic and statistical AI, LLMs, and SLMs.
  • Leverage the AWS ecosystem and modern data platforms to orchestrate secure data migrations and manage multi-stage staging environments.
  • Design and implement cross-platform tools to automate manual curation steps, significantly increasing data velocity.
  • Maintain high-performance pipeline stability and redundancy.
  • Work closely with Sales, Product, and Engineering teams to gather requirements, design solutions, and deliver high-quality data products.
  • Stay current on emerging data technologies and contribute to the evolution of development processes and best practices.
  • Other duties as assigned.

Benefits

medical insurancedental insurancevision insuranceflexible work scheduleunlimited PTO

Skills

AWSCPTCernerDICOMDockerEC2EPICFHIRICD-10IAMLinuxLOINCLLMsOMOPPalantirPythonRDSS3SNOMEDSQLSnowflakeVPC

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free