Lead Data Engineer: Architecting High-Scale Data Solutions
Pelmorex Corp.
About the role
Lead Data Engineer – Transform Environmental Data into Actionable Business Insight
Location: Remote / Flexible (preferably US‑based)
Team: Data & Analytics – Fast‑growing, mission‑driven startup
Why This Role Matters
Our platform turns raw environmental data into powerful, decision‑making intelligence for enterprises worldwide. As a Lead Data Engineer, you’ll design and own the end‑to‑end data fabric that powers this transformation—building the pipelines, orchestration, and APIs that turn massive, heterogeneous data streams into reliable, cost‑effective, real‑time insights.
What You’ll Do
| Area | Impact |
|---|---|
| Architecture & Blueprinting | Define and evolve a scalable, cloud‑native data architecture (lake, warehouse, and serving layers) that can ingest petabytes of sensor, satellite, and IoT data. |
| Pipeline Engineering | Build robust, fault‑tolerant ETL/ELT pipelines using modern stacks (e.g., Apache Beam, Spark, dbt, Airflow/Prefect). Optimize for latency, throughput, and data quality. |
| Workflow Orchestration | Design, implement, and maintain complex DAGs for batch and streaming jobs. Ensure precise scheduling, monitoring, and alerting across environments. |
| API Development | Create high‑performance backend APIs (REST/GraphQL) that expose curated data products to internal tools and external partners. |
| Storage & Cost Optimization | Choose the right storage tier (BigQuery, Cloud Storage, Firestore, etc.) and continuously tune schemas, partitioning, and clustering to balance performance with cost. |
| Mentorship & Best Practices | Lead code‑review cycles, enforce CI/CD pipelines, and champion data‑engineering best practices (testing, documentation, observability). |
| Collaboration | Partner with data scientists, product managers, and business stakeholders to translate requirements into reliable data solutions. |
Who You Are
| Requirement | Details |
|---|---|
| Experience | 5+ years in Data Engineering or Backend Engineering, with a track record of shipping production‑grade data platforms. |
| ETL Mastery | Deep knowledge of ETL/ELT patterns, data modeling, and data quality frameworks. |
| API Expertise | Proven ability to design and deliver performant backend APIs for data consumption. |
| Cloud Proficiency | Hands‑on experience with Google Cloud Platform (BigQuery, Dataflow, Cloud Composer, Pub/Sub, GCS, etc.). Familiarity with IaC (Terraform/Deployment Manager) is a plus. |
| Programming | Strong coding skills in Python, Java/Scala, or Go; comfortable with SQL and scripting. |
| Ops Mindset | Experience with workflow orchestration tools (Airflow, Prefect, Dagster) and monitoring/alerting stacks (Stackdriver, Prometheus, Grafana). |
| Culture Fit | Entrepreneurial spirit, curiosity about environmental data, and a commitment to engineering excellence. |
What We Offer
- Impact‑First Mission – Directly influence sustainability‑focused products used by Fortune‑500 customers.
- Tech‑First Environment – Work with the latest cloud services, open‑source frameworks, and a modern CI/CD pipeline.
- Competitive Compensation – Base salary + equity + performance bonuses.
- Flexibility – Remote‑first culture with flexible hours and generous PTO.
- Learning & Growth – Budget for conferences, certifications, and continuous education.
- Collaborative Culture – Small, cross‑functional team where every voice matters.
Ready to Shape the Future of Data Engineering?
If you’re excited about building world‑class data systems that drive real‑world environmental impact, we’d love to hear from you.
Apply now with your resume, a brief cover letter highlighting a data pipeline you’re proud of, and any relevant GitHub/portfolio links.
Join us and turn data into decisive action.
Requirements
- Over 5 years in Data Engineering or Backend Engineering
- Expertise in ETL patterns and data optimization
- Proven experience in designing backend APIs
- Familiarity with cloud environments, especially GCP
- Strong commitment to engineering best practices
Responsibilities
- Design blueprints for scalable data architectures
- Engineer robust pipelines with modern tech stacks
- Operate complex workflow scheduling with precision
- Create backend APIs delivering reliable data
- Fine-tune storage solutions for performance and cost
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free