Skip to content
mimi

Data Engineer

Network IT

Johannesburg · On-site Full-time Senior 2w ago

About the role

About

Where Big Data Becomes Big Impact. If you believe data should drive solutions, not just storage — this opportunity is built for you. Step into a role where robust engineering meets real‑world business outcomes, and where your ability to build powerful, scalable data ecosystems shapes the future of decision‑making. This is your chance to design, automate, and optimise data pipelines that don’t just support the business… They accelerate it.

Responsibilities

The Mission

Join a high‑performance data engineering team responsible for building modern, reliable, and secure data pipelines across cloud and on‑premises environments. You’ll architect solutions, streamline data flows, integrate diverse sources, and enable analytics teams with structured, accessible, and high‑quality datasets. Working in an agile environment, you’ll be hands‑on in the full engineering lifecycle: designing, developing, deploying, testing, automating, and supporting enterprise‑scale data solutions that fuel strategic insights.

What You’ll Be Building

  • Architect scalable data analytics frameworks across cloud and hybrid environments
  • Translate complex requirements into well‑designed, high‑performing data architecture
  • Lead the development of batch and real‑time data processing solutions
  • Drive the analysis, design, build, testing, and automation of data engineering components
  • Develop and support data feeds from on‑premise systems into AWS
  • Create and manage data marts using Talend or equivalent ETL technologies
  • Process and manipulate large datasets using Python, Spark, and big‑data tooling
  • Work with Hadoop/EMR for distributed compute workloads
  • Build robust, secure, repeatable data pipelines designed for scale
  • Perform database development, documentation, standards implementation, and optimisation
  • Monitor and support production data feeds and troubleshoot break‑fix scenarios
  • Conduct research on emerging data engineering technologies and apply them where beneficial
  • Contribute to business continuity, disaster recovery, and platform stability initiatives

Requirements

Your Technical Toolkit

  • Python (advanced data manipulation and automation)
  • Talend or similar ETL development tools
  • AWS stack: EMR, EC2, S3
  • Hadoop/EMR processing frameworks
  • Spark / PySpark for distributed data workflows
  • Strong SQL and database optimisation skills
  • Expertise in data modelling, distributed systems, and big‑data architecture
  • Experience with automated testing and CI/CD for data workloads

Who You Are

  • Bachelor’s Degree in Computer Science, Computer Engineering, or equivalent
  • AWS Certification (advantageous / preferred)
  • 5+ years’ experience in Data Engineering or Software Engineering
  • 2+ years in Big Data environments
  • 2+ years ETL development experience
  • 2+ years hands‑on AWS platform experience
  • Proven ability to design and support large‑scale data infrastructures
  • Strong understanding of physical database design and SDLC
  • Broad software engineering knowledge (object‑oriented design, coding principles, testing patterns)

Additional Information

For more exciting IT vacancies, please visit: https://www.networkrecruitmentinternational.com/it

I also specialise in recruiting in the following:

  • Data Engineering & Analytics
  • Cloud, DevOps & Infrastructure
  • Software Development (Java, .NET, Python)

If you have not had any response in two weeks, please consider the vacancy application unsuccessful. Your profile will be kept on our database for any other suitable roles / positions.

For more information, contact: Ally-Jane Volmink Specialist Recruitment Consultant Connect with me on LinkedIn! www.linkedin.com/in/ally-jane-volmink-19185925b

Requirements

  • Proven ability to design and support large-scale data infrastructures
  • Strong understanding of physical database design and SDLC
  • Broad software engineering knowledge (object-oriented design, coding principles, testing patterns)

Responsibilities

  • Architect scalable data analytics frameworks across cloud and hybrid environments
  • Translate complex requirements into well-designed, high-performing data architecture
  • Lead the development of batch and real-time data processing solutions
  • Drive the analysis, design, build, testing, and automation of data engineering components
  • Develop and support data feeds from on-premise systems into AWS
  • Create and manage data marts using Talend or equivalent ETL technologies
  • Process and manipulate large datasets using Python, Spark, and big-data tooling
  • Work with Hadoop/EMR for distributed compute workloads
  • Build robust, secure, repeatable data pipelines designed for scale
  • Perform database development, documentation, standards implementation, and optimisation
  • Monitor and support production data feeds and troubleshoot break-fix scenarios
  • Conduct research on emerging data engineering technologies and apply them where beneficial
  • Contribute to business continuity, disaster recovery, and platform stability initiatives

Skills

AWSAWS EMRAWS EC2AWS S3CI/CDData modelingDatabase optimizationETLHadoopPythonPySparkSQLSparkTalendTesting

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free