Data Engineer
Chewy
About the role
Company Overview:
Chewy is seeking a highly experienced and forward-thinking Data Engineer to join our Enterprise Data Systems team, which drives reporting, analytics, and data science capabilities across the organization. In this key role, you will design and build a state-of-the-art data platform that serves as the foundation for data processing and analysis. This position offers the opportunity to provide strategic leadership, mentor team members, and shape the future of Chewy's data infrastructure while supporting our mission to deliver exceptional experiences for pet parents and partners.
Role and Responsibilities:
As a Data Engineer, you will architect and implement scalable data solutions while guiding the data engineering team in best practices and technical excellence. You will collaborate with cross-functional teams to build platforms that support evolving data needs and drive innovation through emerging technologies. Key responsibilities include: • Lead design, development, and optimization of scalable data ingestion and processing frameworks using Spark and Spark Streaming. • Architect and manage data pipelines using Kafka, Kinesis, Airflow, and AWS services including Glue, EMR, S3, SQS, SNS, and Step Functions. • Guide and mentor the data engineering team, establishing best practices and technical guidelines. • Collaborate with data science and analytics teams to design platforms supporting organizational data needs. • Ensure data quality, observability, and governance through implementation of auditing, validation, and monitoring practices. • Work with data warehouse platforms such as Snowflake, Redshift, DynamoDB, and PostgreSQL. • Develop and manage orchestration pipelines using Airflow to optimize job scheduling and operational efficiency. • Evaluate and implement new data tools, technologies, and frameworks to enhance infrastructure. • Implement and maintain CI/CD pipelines and infrastructure-as-code using Terraform. • Drive cloud architecture decisions leveraging AWS services for large-scale data processing. • Partner with stakeholders to understand data needs and deliver reliable, high-quality data products.
Required Skills and Experience: • Proven experience as Lead Data Engineer or Senior Data Engineer in a data-driven environment. • Hands-on expertise designing and developing data pipelines using Python, Spark, Spark Streaming, and Kafka. • Proficiency with cloud data platforms including AWS Glue, EMR, S3, SQS, SNS, and Step Functions. • Strong working knowledge of Snowflake, Redshift, DynamoDB, PostgreSQL, and other data warehouses. • Experience with Airflow for orchestration and automation of data workflows. • Deep understanding of data quality, governance, and observability principles. • Knowledge of CI/CD pipelines and Terraform for cloud infrastructure management. • Strong knowledge of data platforms such as AWS Data Platform, Databricks, or Cloudera. • Excellent problem-solving and leadership skills with ability to work in fast-paced, dynamic environments.
Preferred Qualifications: • Familiarity with Databricks or Cloudera platforms. • Experience setting up and deploying data infrastructure using Terraform. • Previous experience working with CI/CD pipelines to streamline deployment processes.
Compensation and Benefits: • Opportunity to shape the future of data infrastructure within a high-growth organization. • Work with cutting-edge tools and technologies in cloud data engineering. • Lead and mentor a high-performing data engineering team. • Comprehensive benefits package including healthcare coverage and retirement plans. • Professional development and growth opportunities in a collaborative, innovative environment.
About CHEWY:
Chewy is dedicated to transforming the pet care experience by building trusted connections for pet parents and partners. Our Enterprise Data Systems team plays a critical role in enabling data-driven decision-making across the organization through advanced analytics, reporting, and machine learning capabilities. At Chewy, you will be part of a culture that values innovation, collaboration, and continuous learning while making a meaningful impact on the lives of pets and the people who love them. Join us in building the data platform that powers the future of pet care.
Requirements
- Proven experience as Lead Data Engineer or Senior Data Engineer in a data-driven environment.
- Hands-on expertise designing and developing data pipelines using Python, Spark, Spark Streaming, and Kafka.
- Proficiency with cloud data platforms including AWS Glue, EMR, S3, SQS, SNS, and Step Functions.
- Strong working knowledge of Snowflake, Redshift, DynamoDB, PostgreSQL, and other data warehouses.
- Experience with Airflow for orchestration and automation of data workflows.
- Deep understanding of data quality, governance, and observability principles.
- Knowledge of CI/CD pipelines and Terraform for cloud infrastructure management.
- Strong knowledge of data platforms such as AWS Data Platform, Databricks, or Cloudera.
- Excellent problem-solving and leadership skills with ability to work in fast-paced, dynamic environments.
Responsibilities
- Lead design, development, and optimization of scalable data ingestion and processing frameworks using Spark and Spark Streaming.
- Architect and manage data pipelines using Kafka, Kinesis, Airflow, and AWS services including Glue, EMR, S3, SQS, SNS, and Step Functions.
- Guide and mentor the data engineering team, establishing best practices and technical guidelines.
- Collaborate with data science and analytics teams to design platforms supporting organizational data needs.
- Ensure data quality, observability, and governance through implementation of auditing, validation, and monitoring practices.
- Work with data warehouse platforms such as Snowflake, Redshift, DynamoDB, and PostgreSQL.
- Develop and manage orchestration pipelines using Airflow to optimize job scheduling and operational efficiency.
- Evaluate and implement new data tools, technologies, and frameworks to enhance infrastructure.
- Implement and maintain CI/CD pipelines and infrastructure-as-code using Terraform.
- Drive cloud architecture decisions leveraging AWS services for large-scale data processing.
- Partner with stakeholders to understand data needs and deliver reliable, high-quality data products.
Benefits
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free