Skip to content
mimi

R2R Data Engineer(R2R (Record to Report)

Smart Folks Inc.

Austin · On-site Contract Senior Today

About the role

Summary

The R2R Data Engineer is a technical expert responsible for architecting and implementing enterprise-scale data infrastructure that powers critical financial analytics and reporting for A Finance. This role requires deep technical expertise in modern data engineering practices, including building high-performance ETL/ELT pipelines, designing scalable data models, and implementing robust data quality frameworks that ensure accuracy and consistency across financial systems.

This position demands hands-on experience with cloud-native data platforms, advanced SQL optimization, and programmatic data transformations. The engineer will work cross-functionally with business users, FDT, IS&T, data scientists, and other engineers to develop production-grade data services that support financial close processes, regulatory reporting, and strategic decision-making.

You will be working in enterprise data warehouse (Snowflake), Dataiku and lakehouse environments (AWS S3) to design dimensional models, implement data governance policies, and optimize query performance for large-scale financial datasets.

Responsibilities

  • Design and implement scalable data architectures and dimensional models (star/snowflake schemas) that support financial reporting, analytics, and machine learning use cases
  • Develop, test, deploy, monitor, document and troubleshoot complex data pipelines using modern orchestration frameworks with proper error handling, logging, and alerting mechanisms
  • Build and maintain RESTful APIs and microservices for data access and integration with downstream applications (e.g.: Blackline)
  • Implement data quality frameworks including automated validation, reconciliation logic, and anomaly detection to ensure financial data accuracy
  • Optimize SQL queries and data models for performance in Snowflake, including leveraging clustering keys, materialized views, and query optimization techniques
  • Design and implement secure data pipelines with end-to-end encryption, role-based access controls, and compliance with data privacy regulations
  • Collaborate with data scientists and ML engineers to build feature stores and data pipelines that support machine learning model training and inference
  • Establish and enforce data engineering best practices including code reviews, testing strategies (unit, integration, data quality tests), and documentation standards
  • Evaluate and implement emerging technologies in the data engineering space (e.g.: streaming platforms, data quality tools, metadata management solutions)
  • Participate in on-call rotation to support production data pipelines and resolve critical incidents

Key Qualifications

Required Technical Skills:

  • 5+ years of advanced Python programming experience including object-oriented design, asynchronous programming, and package development
  • Expert-level SQL skills including complex joins, window functions, CTEs, query optimization, and performance tuning in databases
  • Hands-on experience designing and implementing data models in Snowflake including time-travel, zero-copy cloning, data sharing, and cost optimization strategies
  • Proven experience building production-grade ETL/ELT pipelines processing large volumes of data
  • Strong experience with AWS services including S3, Lambda, EC2, IAM, Secrets Manager, and CloudWatch
  • Experience implementing data security controls including encryption at rest/in transit, data masking, tokenization, and row-level security
  • Hands-on experience with CI/CD pipelines using GitHub
  • Strong Git version control skills including branching strategies, pull requests, and code review processes
  • Proficiency in shell scripting (Bash) for automation and system administration tasks

Preferred Technical Skills:

  • Experience with streaming data platforms (Kafka, Kinesis, Pub/Sub) and real-time data processing frameworks (Spark Streaming, Flink)
  • Knowledge of containerization (Docker) and orchestration platforms (Kubernetes, ECS)
  • Experience with data catalog and metadata management tools (Alation, Collibra, DataHub)
  • Experience with data quality frameworks (Great Expectations, Soda, Monte Carlo)
  • Experience building and consuming RESTful APIs using frameworks like FastAPI or Flask

Business & Soft Skills:

  • Understanding of financial processes including Record-to-Report (R2R), Order-to-Cash (O2C), Procure-to-Pay (P2P), or financial planning
  • Experience working with ERP systems (SAP, Oracle Financials) and extracting data from these platforms
  • Strong problem-solving skills with ability to debug complex data issues and performance bottlenecks
  • Excellent communication skills with ability to explain technical concepts to non-technical stakeholders
  • Experience working in Agile/Scrum environments with cross-functional teams

Education and Experience

  • Bachelor''s degree in Computer Science, Computer Engineering, Data Engineering, Mathematics, Statistics, or other quantitative discipline required
  • 5+ years of professional experience in data engineering roles with demonstrated expertise in building production data systems
  • Master''s degree in related field preferred

Skills

AWS CloudWatchAWS EC2AWS IAMAWS LambdaAWS S3AWS Secrets ManagerBashCI/CDDataikuDockerECSETLFastAPIFlinkFlaskGitGitHubGreat ExpectationsKafkaKinesisKubernetesMonte CarloOracle FinancialsPostgreSQLPub/SubPythonSAPSnowflakeSodaSpark StreamingSQL

Don't send a generic resume

Paste this job description into Mimi and get a resume tailored to exactly what the hiring team is looking for.

Get started free