Skip to content
mimi

Senior Data Engineer

Golden Opportunities

India · On-site Full-time Senior Yesterday

About the role

About

We are seeking a highly skilled Senior Data Engineer to join our team in Bangalore. The ideal candidate will have expertise in Python, AI, GCP, and AWS, with a strong background in data engineering and machine learning. The successful candidate will be responsible for designing, developing, and maintaining large‑scale data pipelines and cloud‑native applications.

Responsibilities

  • Design, develop, and maintain large‑scale data pipelines using Python and GCP/AWS.
  • Develop and implement ETL processes to extract, transform, and load data from various sources.
  • Collaborate with cross‑functional teams to integrate machine learning models into cloud‑native applications.
  • Develop and maintain APIs using Python to support data exchange between systems.
  • Design and implement data storage solutions using GCP/AWS services such as Bigtable, Cloud Storage, and Cloud SQL.
  • Develop and maintain data processing workflows using GCP/AWS services such as Dataflow, Cloud Functions, and Cloud Pub/Sub.
  • Collaborate with data scientists to integrate generative AI/LLM models into data pipelines and applications.
  • Develop and maintain documentation for data pipelines, APIs, and cloud‑native applications.

Requirements

  • 3‑5 years of experience in data engineering with a focus on Python, AI, GCP, and AWS.
  • Strong understanding of data engineering principles, including data modeling, data warehousing, and data governance.
  • Experience with ETL processes, data pipelines, and data processing workflows.
  • Proficiency in Python programming language, including experience with popular libraries such as NumPy, pandas, and scikit‑learn.
  • Experience with GCP and/or AWS cloud platforms, including services such as Bigtable, Cloud Storage, Cloud SQL, Dataflow, Cloud Functions, and Cloud Pub/Sub.
  • Experience with API development using Python, including frameworks such as Flask and Django.
  • Strong understanding of machine learning concepts, including supervised and unsupervised learning, regression, classification, and clustering.
  • Experience with generative AI/LLM models, including libraries such as TensorFlow and PyTorch.
  • Strong communication and collaboration skills, with the ability to work effectively with cross‑functional teams.

Benefits

  • Medical insurance for self and family members.
  • Free meals and snacks in the office.
  • Access to state‑of‑the‑art technology and tools.
  • Opportunities for professional growth and development.

Skills Required

  • API Development (Python)
  • AWS Cloud
  • Cloud‑Native Applications
  • Data Pipeline Development
  • ETL/Data Engineering
  • GCP (Google Cloud Platform)
  • Generative AI / GenAI
  • LLM (Large Language Models)
  • Machine Learning / AI Integration
  • Python Data

Requirements

  • 3-5 years of experience in data engineering with a focus on Python, AI, GCP, and AWS.
  • Strong understanding of data engineering principles, including data modeling, data warehousing, and data governance.
  • Experience with ETL processes, data pipelines, and data processing workflows.
  • Proficiency in Python programming language, including experience with popular libraries such as NumPy, pandas, and scikit-learn.
  • Experience with GCP and/or AWS cloud platforms, including services such as Bigtable, Cloud Storage, Cloud SQL, Dataflow, Cloud Functions, and Cloud Pub/Sub.
  • Experience with API development using Python, including experience with frameworks such as Flask and Django.
  • Strong understanding of machine learning concepts, including supervised and unsupervised learning, regression, classification, and clustering.
  • Experience with generative AI/LLM models, including experience with popular libraries such as TensorFlow and PyTorch.
  • Strong communication and collaboration skills, with the ability to work effectively with cross-functional teams.

Responsibilities

  • Design, develop, and maintain large-scale data pipelines using Python and GCP/AWS.
  • Develop and implement ETL processes to extract, transform, and load data from various sources.
  • Collaborate with cross-functional teams to integrate machine learning models into cloud-native applications.
  • Develop and maintain APIs using Python to support data exchange between systems.
  • Design and implement data storage solutions using GCP/AWS services such as Bigtable, Cloud Storage, and Cloud SQL.
  • Develop and maintain data processing workflows using GCP/AWS services such as Dataflow, Cloud Functions, and Cloud Pub/Sub.
  • Collaborate with data scientists to integrate generative AI/LLM models into data pipelines and applications.
  • Develop and maintain documentation for data pipelines, APIs, and cloud-native applications.

Benefits

medical insurancemeals and snacksaccess to state-of-the-art technology and toolsopportunities for professional growth and development

Skills

AIAPI DEVELOPMENTAWSCloud-native applicationsData engineeringData pipeline developmentETLGCPGenerative AILLMMachine learningPython

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free