Skip to content
mimi

Scala Data Engineer

SysMind

On-site Full-time 1w ago

About the role

Company Overview

Sysmind Tech PVT LTD is a leading technology solutions provider specializing in data engineering and cloud-based services. We empower businesses across various sectors, including finance, healthcare, and e-commerce, to unlock the value of their data through innovative and scalable solutions. Our expertise lies in building robust data pipelines, data warehouses, and advanced analytics platforms that drive informed decision-making and business growth. We foster a collaborative and growth-oriented culture, encouraging our employees to push boundaries and contribute to cutting-edge projects.

Role Overview

As a Scala Data Engineer at Sysmind Tech, you will be instrumental in designing, developing, and maintaining our data infrastructure. You will work closely with data scientists, analysts, and other engineers to build scalable and reliable data pipelines that ingest, process, and transform large datasets. Your contributions will directly impact the ability of our clients to gain actionable insights from their data, enabling them to optimize their operations and improve their business outcomes.

Key Responsibilities

  • Design and implement scalable and robust data pipelines using Scala and Spark for efficient data processing and storage.
  • Develop and maintain ETL processes to ingest data from various sources into our data warehouse, ensuring data quality and consistency.
  • Build and maintain data models and schemas to support analytical and reporting requirements, enabling effective data analysis.
  • Implement data quality checks and monitoring systems to ensure data accuracy and reliability, improving the trustworthiness of our data assets.
  • Collaborate with data scientists and analysts to understand their data needs and provide them with the necessary data infrastructure and tools, empowering them to perform advanced analytics.
  • Perform ETL/Datawarehouse testing to ensure the accuracy and reliability of data transformations and data loading processes, maintaining data integrity.
  • Deploy and manage data pipelines on AWS cloud infrastructure, ensuring scalability, reliability, and cost-effectiveness.

Required Skillset

  • Demonstrated ability to develop and maintain data pipelines using Scala and Spark for large-scale data processing.
  • Proven expertise in designing and implementing ETL processes for data ingestion, transformation, and loading.
  • Strong understanding of data modeling principles and experience in building data warehouses and data lakes.
  • Proficiency in writing complex SQL queries for data extraction, manipulation, and analysis.
  • Experience with AWS cloud services, including S3, EC2, EMR, and Redshift, for deploying and managing data infrastructure.
  • Excellent communication and collaboration skills to work effectively with cross-functional teams.
  • Bachelor's or Master's degree in Computer Science or a related field.

Requirements

  • Demonstrated ability to develop and maintain data pipelines using Scala and Spark for large-scale data processing.
  • Proven expertise in designing and implementing ETL processes for data ingestion, transformation, and loading.
  • Strong understanding of data modeling principles and experience in building data warehouses and data lakes.
  • Proficiency in writing complex SQL queries for data extraction, manipulation, and analysis.
  • Experience with AWS cloud services, including S3, EC2, EMR, and Redshift, for deploying and managing data infrastructure.
  • Excellent communication and collaboration skills to work effectively with cross-functional teams.

Responsibilities

  • Design and implement scalable and robust data pipelines using Scala and Spark for efficient data processing and storage.
  • Develop and maintain ETL processes to ingest data from various sources into our data warehouse, ensuring data quality and consistency.
  • Build and maintain data models and schemas to support analytical and reporting requirements, enabling effective data analysis.
  • Implement data quality checks and monitoring systems to ensure data accuracy and reliability, improving the trustworthiness of our data assets.
  • Collaborate with data scientists and analysts to understand their data needs and provide them with the necessary data infrastructure and tools, empowering them to perform advanced analytics.
  • Perform ETL/Datawarehouse testing to ensure the accuracy and reliability of data transformations and data loading processes, maintaining data integrity.
  • Deploy and manage data pipelines on AWS cloud infrastructure, ensuring scalability, reliability, and cost-effectiveness.

Skills

AWSAWS EC2AWS EMRAWS RedshiftAWS S3Data modelingData warehousingETLJavaSQLScalaSpark

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free