IS
Solution Architect 15+ Years
Ibroad Solutions
Hyderabad · On-site Full-time Senior Yesterday
About the role
Solution Architect – Data Lake / Big Data
Experience: 15+ Years
Location: Hyderabad (Onsite)
About the Role
We are looking for an experienced Solution Architect with deep expertise in on-premise Data Lake and Big Data ecosystems. The ideal candidate will be responsible for designing, implementing, and optimizing scalable data platforms, ensuring robust architecture, governance, and performance.
Key Responsibilities
- Design and implement on-premise Data Lake architectures and Big Data solutions.
- Define end-to-end data architecture from ingestion to consumption layers.
- Lead the development of scalable ETL/ELT pipelines using Talend and Spark (PySpark).
- Architect solutions using Hadoop ecosystem tools such as HDFS, Hive, Kafka, and Trino/Presto.
- Implement and manage Lakehouse architectures using Apache Hudi, Delta Lake, or Apache Iceberg.
- Ensure data reliability, performance, and scalability through distributed computing principles.
- Collaborate with stakeholders to translate business requirements into technical solutions.
- Drive best practices in data modeling, partitioning, and schema evolution.
Core Technical Skills
Big Data & Data Lake
- Strong experience with Hadoop ecosystem (HDFS, Hive, Spark, Kafka)
- Hands‑on expertise in PySpark and Talend (ETL)
- Experience with Trino/Presto for data querying
- Knowledge of ACID‑compliant lakehouse formats (Hudi, Delta Lake, Iceberg)
Data Engineering & Modeling
- Expertise in data modeling (Raw → Curated → Consumption layers)
- Strong understanding of:
- Partitioning & bucketing
- Schema evolution
- Performance tuning
- Knowledge of distributed systems and parallel processing
Security & Governance
- Experience implementing:
- Kerberos authentication
- Ranger/Sentry authorization frameworks
- LDAP/Active Directory integration
- Strong understanding of:
- TLS/SSL encryption
- Data masking & encryption
- Key management practices
- Familiarity with metadata and governance tools like:
- Apache Atlas / Collibra (or equivalent)
DevOps & Orchestration
- Hands‑on experience with:
- Airflow / Oozie / Control‑M
- Git and CI/CD pipelines
- Knowledge of:
- Monitoring tools such as Grafana, Prometheus, ELK stack
- Experience in environment management and release processes
Good to Have
- Exposure to cloud platforms (Azure / AWS / GCP) and hybrid architectures
- Familiarity with containerization tools (Docker, Kubernetes)
- Understanding of integrating on‑prem Data Lakes with cloud ecosystems
What We’re Looking For
- 15+ years of overall IT experience with strong architectural expertise
- Proven track record in delivering enterprise‑scale Big Data solutions
- Strong problem‑solving and stakeholder management skills
- Ability to lead technical discussions and mentor engineering teams
Job Type
Full‑time
Work Location
In person
Requirements
- 15+ years of overall IT experience with strong architectural expertise
- Proven track record in delivering enterprise-scale Big Data solutions
- Strong problem-solving and stakeholder management skills
- Ability to lead technical discussions and mentor engineering teams
Responsibilities
- Design and implement on-premise Data Lake architectures and Big Data solutions.
- Define end-to-end data architecture from ingestion to consumption layers.
- Lead the development of scalable ETL/ELT pipelines using Talend and Spark (PySpark).
- Architect solutions using Hadoop ecosystem tools such as HDFS, Hive, Kafka, and Trino/Presto.
- Implement and manage Lakehouse architectures using Apache Hudi, Delta Lake, or Apache Iceberg.
- Ensure data reliability, performance, and scalability through distributed computing principles.
- Collaborate with stakeholders to translate business requirements into technical solutions.
- Drive best practices in data modeling, partitioning, and schema evolution.
Skills
Active DirectoryAirflowApache AtlasApache HudiAWSCollibraControl-MDelta LakeDockerELK stackGCPGitGrafanaHadoopHDFSHiveIcebergKafkaKerberosKubernetesLDAPOoziePrometheusRangerSentrySparkTalendTLS/SSLTrinoPresto
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free