Machine Learning Engineering
Qualcomm
About the role
About
As a member of the Low Power AI Solutions team, you will play a critical role in enabling efficient deployment of AI models on Qualcomm's low-power AI accelerators. This position focuses on developing and optimizing the machine learning runtime framework for inference workloads on embedded edge devices. You will be responsible for implementing performance-critical components of the machine learning runtime framework and applying advanced optimization techniques. This role includes adding runtime support for popular ML architectures that are best suited for Qualcomm’s low-power AI accelerators. Your work will directly impact the runtime efficiency, latency, and power consumption of AI applications running on Qualcomm hardware. New Headcount.
Key Responsibilities
- Design and implement core components of the ML runtime framework for inference on embedded systems.
- Collaborate with compiler, hardware, and model teams to co-design efficient execution paths for AI workloads.
- Develop and maintain C/C++ code for runtime kernels and system-level integration.
- Develop tools to assist with performance profiling and debugging of quantized model accuracy.
- Analyze and improve runtime behavior using profiling tools and hardware counters.
- Support deployment of models from popular ML frameworks (e.g., ONNX, TensorFlow, PyTorch) onto Qualcomm’s inference stack.
Required Skills & Experience
- Strong hands‑on experience in performance optimization for embedded or low‑power systems.
- Proficient in C/C++ programming, with a focus on system‑level and runtime development.
- Solid understanding of embedded system design, including memory hierarchy and hardware‑software interaction.
- Experience with Linux/Android development environments and toolchains.
- Familiarity with computer architecture, especially for AI accelerators or DSPs.
- Basic knowledge of machine learning concepts and model structures.
Preferred Qualifications
- Master’s degree in Computer Science, Engineering, or related field.
- 2+ years of experience with ML frameworks (e.g., TensorFlow, PyTorch, ONNX).
- 2+ years of experience in embedded system development and optimization for ML inference.
- 2+ years of experience with C/C++ in performance‑critical environments.
- Experience with low‑level OS interactions (Linux, Android, QNX).
- Familiarity with quantization, graph optimization, and model deployment pipelines.
- Experience working in cross‑functional teams and large matrixed organizations.
Minimum Qualifications
- Option 1: Bachelor's degree in Computer Science, Engineering, Information Systems, or related field and 2+ years of Hardware Engineering, Software Engineering, Systems Engineering, or related work experience.
- Option 2: Master's degree in Computer Science, Engineering, Information Systems, or related field and 1+ year of Hardware Engineering, Software Engineering, Systems Engineering, or related work experience.
- Option 3: PhD in Computer Science, Engineering, Information Systems, or related field.
Pay Range and Other Compensation & Benefits
- Salary: $114,400.00 – $164,400.00 (minimum to maximum pay scale for this location).
- Competitive annual discretionary bonus program.
- Opportunity for annual RSU grants (employees on sales‑incentive plans are not eligible for the annual bonus).
- Highly competitive benefits package designed to support success at work, at home, and at play.
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free