Skip to content
mimi

Cloud Data Engineer

ProcDNA

India · On-site Full-time Mid Level Today

About the role

About ProcDNA

ProcDNA is a global consulting firm. We fuse design thinking with cutting-edge technology to create game-changing Commercial Analytics and Technology solutions for our clients. We're a passionate team of 400+ across 6 offices, all growing and learning together since our launch during the pandemic. Here, you won't be stuck in a cubicle - you'll be out in the open water, shaping the future with brilliant minds. At ProcDNA, innovation isn't just encouraged; it's ingrained in our DNA. Ready to join our epic growth journey?

What we are looking for

Looking for a motivated Cloud Data Engineer with 2–4 years of experience to design, develop, and maintain scalable data pipelines and cloud-based platforms. Hands-on experience with AWS or Azure, along with strong programming skills in Python and SQL, and a solid understanding of ETL/ELT processes, data warehousing, and data modeling is essential. Experience with orchestration, exposure to batch or real-time processing, and familiarity with CI/CD, Git, and basic DevOps/DataOps practices is desirable. Strong communication skills, a collaborative mindset, and the ability to work effectively with cross-functional teams are key expectations.

What you'll do

  • Design, develop, and maintain scalable ETL/ELT pipelines using Databricks and cloud platforms (AWS/Azure).
  • Implement data transformation logic, validation rules, and data quality checks to ensure accuracy, consistency, and reliability of datasets.
  • Develop and manage data models, schemas, and data warehouse structures to enable seamless data consumption for analytics and reporting.
  • Monitor, troubleshoot, and continuously optimize data pipelines for performance, scalability, reliability, and cost efficiency in cloud environments.
  • Ensure adherence to data governance, security, and compliance standards across all data processing and storage layers.
  • Collaborate closely with cross-functional teams to deliver scalable, high-quality, and business-aligned data solutions.

Must Have

  • Minimum 2-4 years of experience in a Data Engineering role with a B.Tech/BE or equivalent technical background.
  • Strong programming skills in Python and SQL with experience in building and optimizing data pipelines.
  • Hands-on experience building ETL/ELT pipelines and cloud data services (Azure or AWS).
  • Experience working with Databricks or similar distributed data processing frameworks (e.g., AWS Glue).
  • Familiarity with workflow orchestration tools like Airflow and version control systems such as Git.
  • Experience handling large datasets, implementing data quality checks, and understanding data ingestion, storage, and consumption architectures.
  • Good understanding of data quality, monitoring, and basic DevOps/DataOps practices, along with strong problem-solving and communication skills.

Requirements

  • Strong programming skills in Python and SQL with experience in building and optimizing data pipelines.
  • Hands-on experience building ETL/ELT pipelines and cloud data services (Azure or AWS).
  • Experience working with Databricks or similar distributed data processing frameworks (e.g., AWS Glue).
  • Familiarity with workflow orchestration tools like Airflow and version control systems such as Git.
  • Experience handling large datasets, implementing data quality checks, and understanding data ingestion, storage, and consumption architectures.
  • Good understanding of data quality, monitoring, and basic DevOps/DataOps practices, along with strong problem-solving and communication skills.

Responsibilities

  • Design, develop, and maintain scalable ETL/ELT pipelines using Databricks and cloud platforms (AWS/Azure).
  • Implement data transformation logic, validation rules, and data quality checks to ensure accuracy, consistency, and reliability of datasets.
  • Develop and manage data models, schemas, and data warehouse structures to enable seamless data consumption for analytics and reporting.
  • Monitor, troubleshoot, and continuously optimize data pipelines for performance, scalability, reliability, and cost efficiency in cloud environments.
  • Ensure adherence to data governance, security, and compliance standards across all data processing and storage layers.
  • Collaborate closely with cross-functional teams to deliver scalable, high-quality, and business-aligned data solutions.

Skills

AWSAWS GlueAzureDatabricksDevOpsETLELTGitSQLPython

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free