C
Senior Data Engineer
CloudDevs
Remote · South Africa Full-time Senior Today
About the role
About
Are you a highly skilled Senior Data Engineer ready to lead and innovate in a dynamic, cloud‑driven environment? We’re looking for a pivotal team member to design, develop, and maintain cutting‑edge data solutions, ensuring scalability, reliability, and top‑tier performance. If you have a strong background in data engineering, a proven track record in leading technical teams, and thrive in an agile setting, we want to hear from you.
Responsibilities
- Building and maintaining efficient ETL/ELT pipelines using tools like Apache Airflow and PySpark.
- Developing robust database schemas, dimensional models (Kimball/Inmon), and supporting data normalisation for both relational and NoSQL databases.
- Contributing to the development and maintenance of our data warehouses, data lakes, and data lakehouses.
- Working with diverse database systems, including Azure SQL, PostgreSQL, Google BigQuery, MongoDB, and Google Firestore.
- Handling structured, semi‑structured, and big data file formats such as Avro, CSV, Parquet, ORC, and Delta.
- Developing and maintaining APIs for seamless data integration and workflows, with a solid understanding of REST and microservices architectures.
- Overseeing codebase maintenance and optimisation, leveraging Git for version control.
- Implementing thorough integration testing and ensuring high‑quality deliverables for all new data processing scenarios.
- Providing technical design and coding assistance to team members, ensuring successful project milestones.
- Assessing and integrating new data sources to meet evolving business needs.
Requirements
- Strong proficiency in Python and SQL (PostgreSQL or SQL Server preferred).
- Hands‑on experience with Apache Airflow and PySpark.
- Familiarity with Databricks is essential.
- Working knowledge of cloud platforms such as Azure, GCP, or AWS.
- Experience with data warehousing concepts, dimensional modelling, and database normalisation.
- Understanding of big data file formats like Avro, Parquet, ORC, and Delta.
- Proficiency in working with APIs, REST, and microservices architectures.
Education & Experience
- A Bachelor’s degree in Computer Science, Data Science, or related fields.
- 5+ years of progressive experience in data engineering, cloud computing, and technology implementation.
- Experience managing multi‑shore projects and working within cloud ecosystems (SaaS/PaaS).
- Proven experience leading technical teams and mentoring team members.
Benefits
- You’ll get to develop your skill set.
- A competitive, industry benchmark compensation.
- Flexible working hours and a remote office setting.
- You’ll be part of a rapidly growing business.
- Work with the absolute masters in the industry and catch some of their energy, vibe, and passion for what we do.
- Great coffee every day, and samoosa Fridays (in‑office of course).
- Plenty of company‑sponsored learning; certifications and incentives.
- Work Hard. Play Hard. Work‑Life Balance.
- No working on your birthday (free day off).
Reference: #J-18808-Ljbffr
Requirements
- Strong proficiency in Python and SQL (PostgreSQL or SQL Server preferred).
- Hands‑on experience with Apache Airflow and PySpark.
- Familiarity with Databricks is essential.
- Working knowledge of cloud platforms such as Azure, GCP, or AWS.
- Experience with data warehousing concepts, dimensional modelling, and database normalisation.
- Understanding of big data file formats like Avro, Parquet, ORC, and Delta.
- Proficiency in working with APIs, REST, and microservices architectures.
Responsibilities
- Building and maintaining efficient ETL/ELT pipelines using tools like Apache Airflow and PySpark.
- Developing robust database schemas, dimensional models (Kimball/Inmon), and supporting data normalisation for both relational and NoSQL databases.
- Contributing to the development and maintenance of our data warehouses, data lakes, and data lakehouses.
- Working with diverse database systems, including Azure SQL, PostgreSQL, Google BigQuery, MongoDB, and Google Firestore.
- Handling structured, semi‑structured, and big data file formats such as Avro, CSV, Parquet, ORC, and Delta.
- Developing and maintaining APIs for seamless data integration and workflows, with a solid understanding of REST and microservices architectures.
- Overseeing codebase maintenance and optimisation, leveraging Git for version control.
- Implementing thorough integration testing and ensuring high‑quality deliverables for all new data processing scenarios.
- Providing technical design and coding assistance to team members, ensuring successful project milestones.
- Assessing and integrating new data sources to meet evolving business needs.
Benefits
flexible working hourscompany-sponsored learningcertificationsincentivespaid day off for birthday
Skills
AWSAPIApache AirflowAzureCSVDatabricksDeltaDockerGitGoogle BigQueryGoogle FirestoreGCPKimballMongoDBNoSQLORCPaaSParquetPostgreSQLPythonPySparkRESTSQLSQL ServerSaaSmicroservices architectures
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free