Skip to content
mimi

Senior Data Engineer (Spark & Python Specialist)

WatersEdge Solutions

Cape Town · On-site Full-time Senior 2w ago

About the role

Location: Remote (South Africa)Employment Type: Full-TimeIndustry: Data Engineering | Cloud Platforms | Financial Services Technology

WatersEdge Solutions is partnering with a client to recruit a highly skilled Senior Data Engineer (Spark & Python Specialist). This is a strong opportunity for a technically advanced engineer who enjoys building scalable, high-performance data solutions in a modern cloud environment. The role is ideal for someone who thrives on optimisation, code-first engineering, and modernising legacy data logic into clean, portable, Python-centric solutions.

About the Role

As a Senior Data Engineer, you’ll serve as a key technical contributor within the engineering team, focused on building, maintaining, and optimising large-scale data processing engines. You’ll work extensively with Spark, PySpark, Delta Lake, and cloud-based lakehouse environments, helping shape a provider-agnostic platform with strong engineering standards, portability, and performance at its core.

Key Responsibilities

• Optimise Spark-based processing through best practices in memory management, shuffle tuning, and partitioning

• Build and maintain data pipelines using Python, PySpark, Delta Lake, and Parquet

• Refactor legacy SQL-based ETL logic into modular, testable, maintainable Python libraries

• Build and optimise medallion architecture layers across Bronze, Silver, and Gold

• Support code-first orchestration approaches using tools such as Airflow, Dagster, or Python-based wrappers

• Participate in code reviews and mentor junior engineers in PySpark best practices

• Contribute to automated testing frameworks using Pytest

• Work closely with data scientists, analysts, and business stakeholders to deliver fit-for-purpose solutions

• Lead initiatives that strengthen data engineering capability, tooling, and technical best practice

• Ensure strong security, compliance, and governance across engineering processes

What You’ll Bring

• Bachelor’s degree in Computer Science, Information Systems, Engineering, or a related field

• 6+ years of Spark / PySpark experience

• Strong production-grade Python capability

• Solid T-SQL skills with experience interpreting and migrating existing SQL logic

• Experience with Azure Synapse Analytics, Dedicated SQL Pools, and Data Factory

• Hands-on expertise with Delta Lake and Parquet in high-volume environments

• Experience with Docker and open-source, cloud-agnostic engineering standards

• Strong collaboration skills and a proven track record modernising large-scale data workloads

• Experience mentoring engineers and contributing to technical excellence across a team

• Strong understanding of security, compliance, and data governance principles

Nice to Have

• Exposure to Microsoft Fabric

• Experience with code-first orchestration tooling such as Airflow or Dagster

• Experience building reusable internal Python libraries and automated testing patterns

• Background in high-scale cloud data platform modernisation

What’s On Offer

• Fully remote role based in South Africa

• Flexible working arrangements

• Wellness support and home office reimbursement

• Continuous learning opportunities

• Competitive salary, ESOP, and recognition for performance

• Supportive and inclusive team culture focused on accountability and work-life balance

Company Culture

This is a team that values transparency, accountability, inclusion, and technical excellence. You’ll join an environment that supports strong engineering standards, continuous learning, and collaborative problem-solving while giving people the flexibility to do their best work in a sustainable way.

If you have not been contacted within 10 working days, please consider your application unsuccessful.

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free