Big Data Engineer – Snowflake, AWS, Spark, Scala
Confidential
About the role
Skills - Snowflake/Hadoop/Scala/Spark/Hive/AWS
Detailed Job Description
Work within an Agile cross functional team to design, develop and maintain data ingestion flows and evolve the platform to orchestrate them.
Design and implement data pipelines, framework and ETL processes.
Ensure data quality, security, and performance. Develop and optimize SQL queries, stored procedures, and views.
Integrate Snowflake with other data sources and BI tools.
Monitor and troubleshoot data jobs and platform issues.
Develop and maintain comprehensive documentation for data pipelines, transformations, and data models.
What do you need to succeed
5+ years of Experience with Big Data technologies used for ETL Hadoop,
5+ Spark, Hive
Well versed with Snowflake Platform.
Strong SQL skills and knowledge of data modeling.
Programming Experience with Scala or Java an API Development.
Experience with ETL tools (e.g., Informatica, Talend, Apache Airflow).
Familiarity with any of the cloud platforms (Preferably AWS).
Knowledge on Gen AI to generate code and improve developer proficiency Knowledge of Python or other scripting languages is a plus.
Excellent problem-solving and communication skills.
Knowledge of SCM, Infrastructure-as-code, and CICD pipelines.
Nice to Have
Experience with workflow management tools like Apache Airflow
Experience with Continuous Integration tools Git Actions, GitHub, Automated Testing tools, Git, or similar tools.
Experience with DockerKubernetesContainersOCP4
Shell Scripting Bachelors or masters degree in computer science, Data Engineering and, or a related field.
Requirements
- 5+ years of Experience with Big Data technologies used for ETL Hadoop,
- 5+ Spark, Hive
- Well versed with Snowflake Platform
- Strong SQL skills and knowledge of data modeling
- Programming Experience with Scala or Java an API Development
- Experience with ETL tools (e.g., Informatica, Talend, Apache Airflow)
- Familiarity with any of the cloud platforms (Preferably AWS)
- Excellent problem-solving and communication skills
- Knowledge of SCM, Infrastructure-as-code, and CICD pipelines
- Experience with workflow management tools like Apache Airflow
- Experience with Continuous Integration tools Git Actions, GitHub, Automated Testing tools, Git, or similar tools
- Experience with DockerKubernetesContainersOCP4
- Shell Scripting Bachelors or masters degree in computer science, Data Engineering and, or a related field
Responsibilities
- Work within an Agile cross functional team to design, develop and maintain data ingestion flows and evolve the platform to orchestrate them
- Design and implement data pipelines, framework and ETL processes
- Ensure data quality, security, and performance
- Develop and optimize SQL queries, stored procedures, and views
- Integrate Snowflake with other data sources and BI tools
- Monitor and troubleshoot data jobs and platform issues
- Develop and maintain comprehensive documentation for data pipelines, transformations, and data models
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free