Skip to content
mimi

Machine Learning Engineering

Qualcomm

Canada · On-site Full-time Senior $114k – $164k/yr Yesterday

About the role

About

As a member of the Low Power AI Solutions team, you will play a critical role in enabling efficient deployment of AI models on Qualcomm's low-power AI accelerators. This position focuses on developing and optimizing the machine learning runtime framework for inference workloads on embedded edge devices. You will be responsible for implementing performance-critical components of the machine learning runtime framework and applying advanced optimization techniques. This role includes adding runtime support for popular ML architectures that are best suited for Qualcomm’s low-power AI accelerators. Your work will directly impact the runtime efficiency, latency, and power consumption of AI applications running on Qualcomm hardware. New Headcount.

Key Responsibilities

  • Design and implement core components of the ML runtime framework for inference on embedded systems.
  • Collaborate with compiler, hardware, and model teams to co-design efficient execution paths for AI workloads.
  • Develop and maintain C/C++ code for runtime kernels and system-level integration.
  • Develop tools to assist with performance profiling and debugging of quantized model accuracy.
  • Analyze and improve runtime behavior using profiling tools and hardware counters.
  • Support deployment of models from popular ML frameworks (e.g., ONNX, TensorFlow, PyTorch) onto Qualcomm’s inference stack.

Required Skills & Experience

  • Strong hands‑on experience in performance optimization for embedded or low‑power systems.
  • Proficient in C/C++ programming, with a focus on system‑level and runtime development.
  • Solid understanding of embedded system design, including memory hierarchy and hardware‑software interaction.
  • Experience with Linux/Android development environments and toolchains.
  • Familiarity with computer architecture, especially for AI accelerators or DSPs.
  • Basic knowledge of machine learning concepts and model structures.

Preferred Qualifications

  • Master’s degree in Computer Science, Engineering, or related field.
  • 2+ years of experience with ML frameworks (e.g., TensorFlow, PyTorch, ONNX).
  • 2+ years of experience in embedded system development and optimization for ML inference.
  • 2+ years of experience with C/C++ in performance‑critical environments.
  • Experience with low‑level OS interactions (Linux, Android, QNX).
  • Familiarity with quantization, graph optimization, and model deployment pipelines.
  • Experience working in cross‑functional teams and large matrixed organizations.

Minimum Qualifications

  • Option 1: Bachelor's degree in Computer Science, Engineering, Information Systems, or related field and 2+ years of Hardware Engineering, Software Engineering, Systems Engineering, or related work experience.
  • Option 2: Master's degree in Computer Science, Engineering, Information Systems, or related field and 1+ year of Hardware Engineering, Software Engineering, Systems Engineering, or related work experience.
  • Option 3: PhD in Computer Science, Engineering, Information Systems, or related field.

Pay Range and Other Compensation & Benefits

  • Salary: $114,400.00 – $164,400.00 (minimum to maximum pay scale for this location).
  • Competitive annual discretionary bonus program.
  • Opportunity for annual RSU grants (employees on sales‑incentive plans are not eligible for the annual bonus).
  • Highly competitive benefits package designed to support success at work, at home, and at play.

Skills

C++CLinuxMLONNXPyTorchTensorFlow

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free