Skip to content
mimi

Solution Architect 15+ Years

Ibroad Solutions

Hyderabad · On-site Full-time Senior Yesterday

About the role

Solution Architect – Data Lake / Big Data

Experience: 15+ Years
Location: Hyderabad (Onsite)

About the Role

We are looking for an experienced Solution Architect with deep expertise in on-premise Data Lake and Big Data ecosystems. The ideal candidate will be responsible for designing, implementing, and optimizing scalable data platforms, ensuring robust architecture, governance, and performance.

Key Responsibilities

  • Design and implement on-premise Data Lake architectures and Big Data solutions.
  • Define end-to-end data architecture from ingestion to consumption layers.
  • Lead the development of scalable ETL/ELT pipelines using Talend and Spark (PySpark).
  • Architect solutions using Hadoop ecosystem tools such as HDFS, Hive, Kafka, and Trino/Presto.
  • Implement and manage Lakehouse architectures using Apache Hudi, Delta Lake, or Apache Iceberg.
  • Ensure data reliability, performance, and scalability through distributed computing principles.
  • Collaborate with stakeholders to translate business requirements into technical solutions.
  • Drive best practices in data modeling, partitioning, and schema evolution.

Core Technical Skills

Big Data & Data Lake

  • Strong experience with Hadoop ecosystem (HDFS, Hive, Spark, Kafka)
  • Hands‑on expertise in PySpark and Talend (ETL)
  • Experience with Trino/Presto for data querying
  • Knowledge of ACID‑compliant lakehouse formats (Hudi, Delta Lake, Iceberg)

Data Engineering & Modeling

  • Expertise in data modeling (Raw → Curated → Consumption layers)
  • Strong understanding of:
    • Partitioning & bucketing
    • Schema evolution
    • Performance tuning
  • Knowledge of distributed systems and parallel processing

Security & Governance

  • Experience implementing:
    • Kerberos authentication
    • Ranger/Sentry authorization frameworks
    • LDAP/Active Directory integration
  • Strong understanding of:
    • TLS/SSL encryption
    • Data masking & encryption
    • Key management practices
  • Familiarity with metadata and governance tools like:
    • Apache Atlas / Collibra (or equivalent)

DevOps & Orchestration

  • Hands‑on experience with:
    • Airflow / Oozie / Control‑M
    • Git and CI/CD pipelines
  • Knowledge of:
    • Monitoring tools such as Grafana, Prometheus, ELK stack
    • Experience in environment management and release processes

Good to Have

  • Exposure to cloud platforms (Azure / AWS / GCP) and hybrid architectures
  • Familiarity with containerization tools (Docker, Kubernetes)
  • Understanding of integrating on‑prem Data Lakes with cloud ecosystems

What We’re Looking For

  • 15+ years of overall IT experience with strong architectural expertise
  • Proven track record in delivering enterprise‑scale Big Data solutions
  • Strong problem‑solving and stakeholder management skills
  • Ability to lead technical discussions and mentor engineering teams

Job Type

Full‑time

Work Location

In person

Requirements

  • 15+ years of overall IT experience with strong architectural expertise
  • Proven track record in delivering enterprise-scale Big Data solutions
  • Strong problem-solving and stakeholder management skills
  • Ability to lead technical discussions and mentor engineering teams

Responsibilities

  • Design and implement on-premise Data Lake architectures and Big Data solutions.
  • Define end-to-end data architecture from ingestion to consumption layers.
  • Lead the development of scalable ETL/ELT pipelines using Talend and Spark (PySpark).
  • Architect solutions using Hadoop ecosystem tools such as HDFS, Hive, Kafka, and Trino/Presto.
  • Implement and manage Lakehouse architectures using Apache Hudi, Delta Lake, or Apache Iceberg.
  • Ensure data reliability, performance, and scalability through distributed computing principles.
  • Collaborate with stakeholders to translate business requirements into technical solutions.
  • Drive best practices in data modeling, partitioning, and schema evolution.

Skills

Active DirectoryAirflowApache AtlasApache HudiAWSCollibraControl-MDelta LakeDockerELK stackGCPGitGrafanaHadoopHDFSHiveIcebergKafkaKerberosKubernetesLDAPOoziePrometheusRangerSentrySparkTalendTLS/SSLTrinoPresto

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free