Skip to content
mimi

Lead Data Engineer / Data Architect (Supply Chain)

Astir IT Solutions, Inc.

Philadelphia · On-site Full-time Lead 2w ago

About the role

Key Responsibilities

Data Architecture & Design

  • Design and develop Supply Chain Anomaly Detection and Revenue Assurance platforms for order processing data.
  • Define and own the end-to-end data architecture, including ingestion, transformation, storage, and consumption layers.
  • Design robust data models for inventory, logistics, fulfillment, and supplier performance domains.
  • Establish architecture standards, design patterns, and best practices aligned with enterprise data strategy.

Data Engineering & Platforms

  • Architect and guide development of scalable data pipelines using:
    • PySpark / Spark-based processing
    • Python for data transformation and orchestration
    • Enterprise ETL/ELT frameworks
    • Advanced SQL for analytics and modeling
  • Support both batch and near real-time processing use cases.
  • Optimize pipelines for data quality, performance, scalability, and cost efficiency.

Supply Chain Analytics Enablement

  • Enable downstream analytics for:
    • Supply chain planning and forecasting
    • Inventory optimization and demand analytics
    • Vendor and procurement performance reporting
    • Operational KPIs and executive dashboards
    • SKU management
  • Partner with analytics and data science teams to ensure data readiness for advanced analytics.

Cloud & Data Storage

  • Design and implement cloud-native data solutions.
  • Ensure secure, scalable, and resilient data storage and access patterns.

Data Governance & Quality

  • Collaborate with governance and security teams to ensure:
    • Data quality, consistency, and reliability
    • Data lineage, metadata management, and documentation
    • Compliance with data privacy, security, and regulatory standards

Leadership & Collaboration

  • Collaborate with product owners, supply chain leaders, engineering teams, and vendors.
  • Translate business and operational requirements into scalable technical solutions.
  • Mentor and guide data engineers and architects on best practices and design principles.

Qualifications

  • Experience: 12+ years in Data Engineering / Data Architecture
  • Data Pipelines: Strong experience with Kafka, CDC, ETL/ELT tools
  • Streaming: Hands-on experience with Spark Streaming, Kafka Streams
  • Programming: Proficiency in Python, PySpark, SQL
  • Analytics: Strong SQL and experience building reports, dashboards, and KPIs
  • Data Governance: Experience with data lineage tools (e.g., OpenLineage) and compliance frameworks
  • Soft Skills: Strong communication and stakeholder management skills
  • Education: Bachelor’s degree in Computer Science, Engineering, or equivalent

Skills

CDCData GovernanceData LineageETLKafkaKafka StreamsMetadata ManagementOpenLineagePythonPySparkSQLSpark Streaming

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free