Data Engineer
Network IT
About the role
About
Where Big Data Becomes Big Impact. If you believe data should drive solutions, not just storage — this opportunity is built for you. Step into a role where robust engineering meets real‑world business outcomes, and where your ability to build powerful, scalable data ecosystems shapes the future of decision‑making. This is your chance to design, automate, and optimise data pipelines that don’t just support the business… They accelerate it.
Responsibilities
The Mission
Join a high‑performance data engineering team responsible for building modern, reliable, and secure data pipelines across cloud and on‑premises environments. You’ll architect solutions, streamline data flows, integrate diverse sources, and enable analytics teams with structured, accessible, and high‑quality datasets. Working in an agile environment, you’ll be hands‑on in the full engineering lifecycle: designing, developing, deploying, testing, automating, and supporting enterprise‑scale data solutions that fuel strategic insights.
What You’ll Be Building
- Architect scalable data analytics frameworks across cloud and hybrid environments
- Translate complex requirements into well‑designed, high‑performing data architecture
- Lead the development of batch and real‑time data processing solutions
- Drive the analysis, design, build, testing, and automation of data engineering components
- Develop and support data feeds from on‑premise systems into AWS
- Create and manage data marts using Talend or equivalent ETL technologies
- Process and manipulate large datasets using Python, Spark, and big‑data tooling
- Work with Hadoop/EMR for distributed compute workloads
- Build robust, secure, repeatable data pipelines designed for scale
- Perform database development, documentation, standards implementation, and optimisation
- Monitor and support production data feeds and troubleshoot break‑fix scenarios
- Conduct research on emerging data engineering technologies and apply them where beneficial
- Contribute to business continuity, disaster recovery, and platform stability initiatives
Requirements
Your Technical Toolkit
- Python (advanced data manipulation and automation)
- Talend or similar ETL development tools
- AWS stack: EMR, EC2, S3
- Hadoop/EMR processing frameworks
- Spark / PySpark for distributed data workflows
- Strong SQL and database optimisation skills
- Expertise in data modelling, distributed systems, and big‑data architecture
- Experience with automated testing and CI/CD for data workloads
Who You Are
- Bachelor’s Degree in Computer Science, Computer Engineering, or equivalent
- AWS Certification (advantageous / preferred)
- 5+ years’ experience in Data Engineering or Software Engineering
- 2+ years in Big Data environments
- 2+ years ETL development experience
- 2+ years hands‑on AWS platform experience
- Proven ability to design and support large‑scale data infrastructures
- Strong understanding of physical database design and SDLC
- Broad software engineering knowledge (object‑oriented design, coding principles, testing patterns)
Additional Information
For more exciting IT vacancies, please visit: https://www.networkrecruitmentinternational.com/it
I also specialise in recruiting in the following:
- Data Engineering & Analytics
- Cloud, DevOps & Infrastructure
- Software Development (Java, .NET, Python)
If you have not had any response in two weeks, please consider the vacancy application unsuccessful. Your profile will be kept on our database for any other suitable roles / positions.
For more information, contact: Ally-Jane Volmink Specialist Recruitment Consultant Connect with me on LinkedIn! www.linkedin.com/in/ally-jane-volmink-19185925b
Requirements
- Proven ability to design and support large-scale data infrastructures
- Strong understanding of physical database design and SDLC
- Broad software engineering knowledge (object-oriented design, coding principles, testing patterns)
Responsibilities
- Architect scalable data analytics frameworks across cloud and hybrid environments
- Translate complex requirements into well-designed, high-performing data architecture
- Lead the development of batch and real-time data processing solutions
- Drive the analysis, design, build, testing, and automation of data engineering components
- Develop and support data feeds from on-premise systems into AWS
- Create and manage data marts using Talend or equivalent ETL technologies
- Process and manipulate large datasets using Python, Spark, and big-data tooling
- Work with Hadoop/EMR for distributed compute workloads
- Build robust, secure, repeatable data pipelines designed for scale
- Perform database development, documentation, standards implementation, and optimisation
- Monitor and support production data feeds and troubleshoot break-fix scenarios
- Conduct research on emerging data engineering technologies and apply them where beneficial
- Contribute to business continuity, disaster recovery, and platform stability initiatives
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free