AWS CDK experience is mandatory in addition to AWS data engineering skills
Ingest data from internal and external sources into AWS Redshift and S3 using DMS, Zero-ETL, Kinesis, Glue, Lambda, Lake Formation, Cross Account Replication, and/or SFTP
Build infrastructure as code using AWS CDK
Create and utilize GitLab CI/CD pipelines to promote code through test and production environments
Build Glue ETL pipelines to structure and curate data
Conduct code reviews and ensure high-quality deployments
Maintain AWS environments to optimize costs, eliminate vulnerabilities, and ensure smooth operation
Coordinate and resolve production issues promptly
Drive path-to-production processes, including documentation and approvals
Partner with business teams on data governance
Minimum 5 years of post-degree professional experience
3+ years experience with AWS CDK and AWS ETL pipelines
Hands-on experience with AWS services such as S3, Lambda, Step Functions, Glue, and IAM
Experience with cloud data migration tools like DMS and Cross Account Replication
Strong knowledge of Python and software development lifecycle (SDLC)
Familiarity with best practices for data ingestion, data design, and query optimization (indexes, materialized views)
Experience in profiling data, validating analysis, and defining deployment paths
Excellent written and verbal communication skills for cross-functional collaboration
AWS CDK, AWS Data Engineering, AWS Redshift, AWS S3, AWS DMS, Zero-ETL, Kinesis, Glue, Lambda, Lake Formation, Cross Account Replication, SFTP, GitLab CI/CD, Infrastructure as Code (IaC), Glue ETL pipelines, Code Reviews, AWS Environment Management, Production Issue Resolution, Data Governance, Python, Software Development Lifecycle (SDLC), Data Ingestion, Data Design, Query Optimization (indexes, materialized views), Data Profiling, Data Validation, Documentation, Communication, Cross-functional Collaboration.