Skip to content
mimi

Big Data Engineer – Snowflake, AWS, Spark, Scala

Confidential

US · On-site Full-time Today

About the role

Skills - Snowflake/Hadoop/Scala/Spark/Hive/AWS

Detailed Job Description

Work within an Agile cross functional team to design, develop and maintain data ingestion flows and evolve the platform to orchestrate them.

Design and implement data pipelines, framework and ETL processes.

Ensure data quality, security, and performance. Develop and optimize SQL queries, stored procedures, and views.

Integrate Snowflake with other data sources and BI tools.

Monitor and troubleshoot data jobs and platform issues.

Develop and maintain comprehensive documentation for data pipelines, transformations, and data models.

What do you need to succeed

5+ years of Experience with Big Data technologies used for ETL Hadoop,

5+ Spark, Hive

Well versed with Snowflake Platform.

Strong SQL skills and knowledge of data modeling.

Programming Experience with Scala or Java an API Development.

Experience with ETL tools (e.g., Informatica, Talend, Apache Airflow).

Familiarity with any of the cloud platforms (Preferably AWS).

Knowledge on Gen AI to generate code and improve developer proficiency Knowledge of Python or other scripting languages is a plus.

Excellent problem-solving and communication skills.

Knowledge of SCM, Infrastructure-as-code, and CICD pipelines.

Nice to Have

Experience with workflow management tools like Apache Airflow

Experience with Continuous Integration tools Git Actions, GitHub, Automated Testing tools, Git, or similar tools.

Experience with DockerKubernetesContainersOCP4

Shell Scripting Bachelors or masters degree in computer science, Data Engineering and, or a related field.

Requirements

  • 5+ years of Experience with Big Data technologies used for ETL Hadoop,
  • 5+ Spark, Hive
  • Well versed with Snowflake Platform
  • Strong SQL skills and knowledge of data modeling
  • Programming Experience with Scala or Java an API Development
  • Experience with ETL tools (e.g., Informatica, Talend, Apache Airflow)
  • Familiarity with any of the cloud platforms (Preferably AWS)
  • Excellent problem-solving and communication skills
  • Knowledge of SCM, Infrastructure-as-code, and CICD pipelines
  • Experience with workflow management tools like Apache Airflow
  • Experience with Continuous Integration tools Git Actions, GitHub, Automated Testing tools, Git, or similar tools
  • Experience with DockerKubernetesContainersOCP4
  • Shell Scripting Bachelors or masters degree in computer science, Data Engineering and, or a related field

Responsibilities

  • Work within an Agile cross functional team to design, develop and maintain data ingestion flows and evolve the platform to orchestrate them
  • Design and implement data pipelines, framework and ETL processes
  • Ensure data quality, security, and performance
  • Develop and optimize SQL queries, stored procedures, and views
  • Integrate Snowflake with other data sources and BI tools
  • Monitor and troubleshoot data jobs and platform issues
  • Develop and maintain comprehensive documentation for data pipelines, transformations, and data models

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free