Skip to content
mimi

Data Engineer (Support)

Augmented Systems LLP

India · On-site Full-time ₹1200k – ₹1500k/yr 1w ago

About the role

Job Overview

We are seeking a skilled and driven Data Engineer with strong expertise in SQL, PySpark, and Databricks to build, optimize, and maintain scalable data pipelines. The ideal candidate will play a key role in designing robust data solutions, ensuring data quality, and enabling data-driven decision-making across the organization.

Key Responsibilities

  • Data Pipeline Development
    Design, build, and maintain scalable data pipelines using PySpark and Databricks for efficient data processing.

  • Data Transformation & Processing
    Develop and optimize complex data transformations using SQL and PySpark to support analytics and reporting needs.

  • ETL Workflow Management
    Implement and enhance ETL processes to ensure efficient data ingestion, transformation, and loading across systems.

  • Data Quality & Validation
    Ensure accuracy, consistency, and reliability of data through validation checks and quality assurance practices.

  • Performance Optimization
    Analyze and optimize data workflows and queries for improved performance and scalability.

  • Collaboration with Stakeholders
    Work closely with data analysts, data scientists, and business teams to understand requirements and deliver data solutions.

  • Data Platform Management
    Utilize Databricks for managing data engineering workflows, job orchestration, and large-scale data processing.

Required Skills & Qualifications

  • Strong proficiency in SQL for data querying and transformation
  • Hands-on experience with PySpark for large-scale data processing
  • Practical experience working with Databricks platform
  • Solid understanding of data engineering concepts, ETL processes, and data pipelines
  • Experience with data modeling and data warehousing concepts
  • Familiarity with distributed computing frameworks

Preferred Skills

  • Experience with cloud platforms (Azure, AWS, or GCP)
  • Knowledge of workflow orchestration tools (e.g., Airflow, Autosys)
  • Exposure to BI tools such as Power BI
  • Basic understanding of Python programming beyond PySpark

Soft Skills

  • Strong analytical and problem-solving skills
  • Effective communication and collaboration abilities
  • Ability to work in a fast-paced, data-driven environment
  • Attention to detail and commitment to data quality

Job Details

  • Job Types: Full-time, Permanent
  • Pay: ₹1,200,000.00 - ₹1,500,000.00 per year
  • Work Location: In person

Requirements

  • Strong proficiency in SQL for data querying and transformation
  • Hands-on experience with PySpark for large-scale data processing
  • Practical experience working with Databricks platform
  • Solid understanding of data engineering concepts, ETL processes, and data pipelines
  • Experience with data modeling and data warehousing concepts
  • Familiarity with distributed computing frameworks

Responsibilities

  • Design, build, and maintain scalable data pipelines using PySpark and Databricks for efficient data processing.
  • Develop and optimize complex data transformations using SQL and PySpark to support analytics and reporting needs.
  • Implement and enhance ETL processes to ensure efficient data ingestion, transformation, and loading across systems.
  • Ensure accuracy, consistency, and reliability of data through validation checks and quality assurance practices.
  • Analyze and optimize data workflows and queries for improved performance and scalability.
  • Work closely with data analysts, data scientists, and business teams to understand requirements and deliver data solutions.
  • Utilize Databricks for managing data engineering workflows, job orchestration, and large-scale data processing.

Skills

DatabricksETLPySparkSQL

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free