Lead Data Engineer
CGI
About the role
About
As a Lead Data Engineer, you will lead a team of data engineers and collaborate with architects, engineers, information analysts, business, and technology stakeholders to develop and deploy enterprise‑grade platforms that enable data‑driven solutions. You will be responsible for updating the design parameters and implementation details of robust data pipelines, optimizing data processes, ensuring data quality, security, and governance. The role requires knowledge of GCP services that include all GCP‑based services for GCP Dataflow, Cloud Run Functions, AlloyDB, Pub/Sub streams, IAC, BigQuery and Data Plex. This is an event‑driven architecture processing real‑time and batch data from legacy external systems.
The position can be located at any CGI office in the U.S. (preferred location: Fairfax, VA); a hybrid working model is acceptable.
Responsibilities
- Lead the design, development, and maintenance of robust data pipelines using Dataflow and Cloud Run Functions that extract data from various sources, transform it, and load it into the appropriate landing zone of Pub/Sub, Google Storage, or AlloyDB for PostgreSQL.
- Collaborate with FAA customers, chief architect, senior engineers, information analysts, business, and technology stakeholders to develop and deploy enterprise‑grade platforms that enable data‑driven solutions.
- Develop and manage ETL (Extract, Transform, Load) processes using GCP Dataflow with GCP flex templates (Java) to support data transformations into XML and JSON documents, enriching them for downstream data consumers.
- Implement and manage advanced data models, including relational databases, non‑relational databases, master data management, Data Plex, and data governance.
- Integrate data from legacy sources (databases, data warehouses, APIs, external systems) using Dataflow and Cloud Run Functions.
- Ensure data consistency and integrity during integration, performing data validation and cleaning as needed, and manage data quality.
- Transform raw data into a usable format by applying cleansing, aggregation, filtering, and enrichment techniques using Dataflow, Cloud Run Functions, and other GCP services.
- Design and optimize data pipelines and processing workflows for performance, scalability, and efficiency using Dataflow and Cloud Run Functions.
- Monitor and tune data systems, identify and resolve performance bottlenecks, and implement indexing strategies to enhance query performance.
- Implement data quality checks and validations within pipelines to ensure accuracy, consistency, and completeness of data.
- Optimize and administer GCP data environments and related services to ensure high performance and reliability.
- Collaborate with cross‑functional teams to deliver data products.
Required Qualifications
- Minimum of 12 years of experience in data engineering, with at least 2 years of hands‑on experience with Google Cloud‑based services.
- At least 8 years of work experience in data solutions design and management, including data integration, modeling, optimization, and data quality.
- Project management skills: able to understand business needs, prioritize, and manage workload effectively.
- Data architect experience: provide high‑level guidance, ensure operational stability, long‑term maintainability, and support 24/7 application data operations.
- Strong knowledge of SQL, Python, and Java (Dataflow flex templates are used with Java).
- Experience with cloud platforms such as GCP, GKE, IAM, Terraform, CI/CD to deliver automated solutions.
- Excellent problem‑solving and debugging skills.
- Strong communication skills and a proactive “getting things done” mindset.
- Experience working in Agile teams; familiarity with Agile methodologies, SCRUM with 2‑week sprints.
- Ability to design, build, and deploy data solutions that capture, explore, transform, and utilize data for mission applications, insights, and reporting.
- Experience with database technologies such as SQL, NoSQL, AlloyDB PostgreSQL.
- Ability to collaborate across teams with differing technical knowledge to support delivery of data products.
- Strong knowledge of data architecture in application development or reporting.
- Excellent organizational and analytical abilities.
- Good written and verbal communication skills.
Desired Qualifications / Non‑essential Skills
- Industry knowledge in sectors such as Government and Aviation preferred.
Compensation
A reasonable estimate of the current range for this role in the U.S. is $113,400 – $223,200.
Benefits
- Competitive compensation
- Comprehensive insurance options
- Matching contributions through the 401(k) plan and the share purchase plan
- Paid time off for vacation, holidays, and sick time
- Paid parental leave
- Learning opportunities and tuition assistance
- Wellness and well‑being programs
Skills
Don't send a generic resume
Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.
Get started free