Skip to content
mimi

Data Engineer

KPMG

Calgary · On-site Full-time Senior CA$73k – CA$102k/yr Today

About the role

About

At KPMG, you’ll join a team of diverse and dedicated problem solvers, connected by a common cause: turning insight into opportunity for clients and communities around the world.

Are you a talented leader with a proven track record for motivating teams and delivering exceptional client service?

Our team is looking for a Data Engineer with extensive hands‑on expertise in Databricks and strong consulting capability. This role will support and lead modernization initiatives from legacy/on‑prem data platforms to scalable, secure, and cost‑optimized Lakehouse architectures using Databricks and similar technologies.

Responsibilities

  • Partner with clients to understand business goals, gather requirements, and translate them into actionable technical designs and delivery plans.
  • Work with the engagement team to translate business and analytics requirements into a data strategy for the engagement including ETL/ELT, data model, and staging data for analysis.
  • Contribute to end‑to‑end solution architecture for repeatable, cost‑optimized implementations (including non‑functional requirements and operational readiness).
  • Lead delivery of modern data platforms on Databricks (ETL/ELT pipelines, workload migrations, governance enablement).
  • Implement Delta Lake / Lakehouse patterns including medallion architecture, CDC, incremental processing, and data quality controls.
  • Develop data pipelines to support streaming, incremental, batch data, etc.
  • Design and implement scalable batch and streaming pipelines using Spark and modern orchestration patterns.
  • Apply CI/CD and engineering best practices (version control, automated deployment, testing, and release management) to data engineering workflows.
  • Establish and operationalize governance using Unity Catalog, including access controls, lineage, and security frameworks.
  • Support testing and production releases, including troubleshooting, performance tuning, and stabilization.
  • Proactively contribute to the creation of presentation materials relating to data activities for stakeholder discussions.

Requirements

  • University degree in computer engineering, mathematics, data science or related disciplines
  • 4+ years of professional experience in a related field like Data Engineering, Business Intelligence, or related field with a track record of manipulating, processing, and extracting value from large datasets.
  • 2+ years of hands‑on experience with Databricks, including advanced features (Delta Lake, Unity Catalog) or cloud certifications, with 1‑2 years of experience leading workstreams / client‑facing delivery.
  • Strong proficiency in SQL and solid understanding of modern data modeling principles, dimensional modeling, and data warehousing concepts.
  • Proficiency in Python (or similar scripting languages) for data processing, automation, and analytical workflows.
  • Strong experience working in teams to perform ETL (extract, transform and load) of data from a variety of databases (SQL, NoSQL, etc.).
  • Proven experience leading large‑scale data migrations (ETL, workloads, cloud platforms), including migration of legacy data platforms or ETL workloads to cloud‑native environments.
  • Experience applying CI/CD practices to data engineering workflows, including version control, automated deployment, and pipeline orchestration.
  • Independent ability to review data quality and data definitions and perform data cleansing and data management tasks.
  • Experience collaborating within cross‑functional and multi‑disciplinary teams to solve complex data challenges, including processing semi‑structured and unstructured data.
  • Experience in at least one major cloud service (AWS, Azure, or GCP) with understanding of cloud‑native services, identity management, and scalable architecture principles.
  • Certifications: Databricks Certified Data Engineer (Associate or Professional) and/or relevant cloud certifications (e.g., Azure, AWS, or GCP architecture or data engineering credentials) are preferred.

Requirements

  • University degree in computer engineering, mathematics, data science or related disciplines
  • 4+ years of professional experience in a related field like Data Engineering, Business Intelligence, or related field with a track record of manipulating, processing, and extracting value from large datasets.
  • 2+ years of hands-on experience with Databricks, including advanced features (Delta Lake, Unity Catalog) with Databricks or cloud certifications with 1-2 years of experience leading workstreams / client-facing delivery.
  • Strong proficiency in SQL and solid understanding of modern data modeling principles, dimensional modeling, and data warehousing concepts.
  • Proficiency in Python (or similar scripting languages) for data processing, automation, and analytical workflows
  • Strong experience working in teams to perform ETL (extract, transform and load) of data from a variety of databases from SQL, NoSQL, etc.
  • Proven experience leading large-scale data migrations (ETL, workloads, cloud platforms), including migration of legacy data platforms or ETL workloads to cloud-native environments.
  • Experience applying CI/CD practices to data engineering workflows, including version control, automated deployment, and pipeline orchestration.
  • Independent ability to review the data quality and data definitions and perform data cleansing and data management tasks.
  • Experience collaborating within cross-functional and multi-disciplinary teams to solve complex data challenges, including processing semi-structured and unstructured data
  • Experience in at least one major cloud service: AWS, Azure and GCP with understanding of cloud-native services, identity management, and scalable architecture principles.

Responsibilities

  • Partner with clients to understand business goals, gather requirements, and translate them into actionable technical designs and delivery plans.
  • Work with the engagement team to translate business and analytics requirements into a data strategy for the engagement including ETL/ELT, data model, and staging data for analysis.
  • Contribute to end-to-end solution architecture for repeatable, cost-optimized implementations (including non-functional requirements and operational readiness).
  • Lead delivery of modern data platforms on Databricks (ETL/ELT pipelines, workload migrations, governance enablement).
  • Implement Delta Lake / Lakehouse patterns including medallion architecture, CDC, incremental processing, and data quality controls.
  • Develop data pipelines to support streaming, incremental, batch data, etc.
  • Design and implement scalable batch and streaming pipelines using Spark and modern orchestration patterns.
  • Apply CI/CD and engineering best practices (version control, automated deployment, testing, and release management) to data engineering workflows.
  • Establish and operationalize governance using Unity Catalog, including access controls, lineage, and security frameworks.
  • Support testing and production releases, including troubleshooting, performance tuning, and stabilization.
  • Proactively contributes to the creation of presentation materials relating to data activities for stakeholder discussions.

Benefits

bonus awardsTotal Rewards program

Skills

AWSAzureDatabricksDelta LakeGCPLakehouseNoSQLPythonSQLSparkUnity Catalogversion control

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free