Data Engineer (Support)
Augmented Systems LLP
About the role
Job Overview
We are seeking a skilled and driven Data Engineer with strong expertise in SQL, PySpark, and Databricks to build, optimize, and maintain scalable data pipelines. The ideal candidate will play a key role in designing robust data solutions, ensuring data quality, and enabling data-driven decision-making across the organization.
Key Responsibilities
Data Pipeline Development
Design, build, and maintain scalable data pipelines using PySpark and Databricks for efficient data processing.Data Transformation & Processing
Develop and optimize complex data transformations using SQL and PySpark to support analytics and reporting needs.ETL Workflow Management
Implement and enhance ETL processes to ensure efficient data ingestion, transformation, and loading across systems.Data Quality & Validation
Ensure accuracy, consistency, and reliability of data through validation checks and quality assurance practices.Performance Optimization
Analyze and optimize data workflows and queries for improved performance and scalability.Collaboration with Stakeholders
Work closely with data analysts, data scientists, and business teams to understand requirements and deliver data solutions.Data Platform Management
Utilize Databricks for managing data engineering workflows, job orchestration, and large-scale data processing.
Required Skills & Qualifications
- Strong proficiency in SQL for data querying and transformation
- Hands-on experience with PySpark for large-scale data processing
- Practical experience working with Databricks platform
- Solid understanding of data engineering concepts, ETL processes, and data pipelines
- Experience with data modeling and data warehousing concepts
- Familiarity with distributed computing frameworks
Preferred Skills
- Experience with cloud platforms (Azure, AWS, or GCP)
- Knowledge of workflow orchestration tools (e.g., Airflow, Autosys)
- Exposure to BI tools such as Power BI
- Basic understanding of Python programming beyond PySpark
Soft Skills
- Strong analytical and problem-solving skills
- Effective communication and collaboration abilities
- Ability to work in a fast-paced, data-driven environment
- Attention to detail and commitment to data quality
Job Details
- Job Types: Full-time, Permanent
- Pay: ₹1,200,000.00 - ₹1,500,000.00 per year
- Work Location: In person
Requirements
- Strong proficiency in SQL for data querying and transformation
- Hands-on experience with PySpark for large-scale data processing
- Practical experience working with Databricks platform
- Solid understanding of data engineering concepts, ETL processes, and data pipelines
- Experience with data modeling and data warehousing concepts
- Familiarity with distributed computing frameworks
Responsibilities
- Design, build, and maintain scalable data pipelines using PySpark and Databricks for efficient data processing.
- Develop and optimize complex data transformations using SQL and PySpark to support analytics and reporting needs.
- Implement and enhance ETL processes to ensure efficient data ingestion, transformation, and loading across systems.
- Ensure accuracy, consistency, and reliability of data through validation checks and quality assurance practices.
- Analyze and optimize data workflows and queries for improved performance and scalability.
- Work closely with data analysts, data scientists, and business teams to understand requirements and deliver data solutions.
- Utilize Databricks for managing data engineering workflows, job orchestration, and large-scale data processing.
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free