AWS Data Engineer with SageMaker
Location- Reston, VA (In person Client interview) day one onsite
JD :
Seeking an AWS Data Engineer to design, build, and maintain scalable data pipelines and ETL solutions using Python / Pyspark and AWS managed services to support analytics and data product needs.
Key Responsibilities
- Build and maintain ETL pipelines using Python and PySpark on AWS Glue and other compute platforms
- Orchestrate workflows with AWS Step Functions and serverless components ( Lambda )
- Implement messaging and event-driven patterns using AWS SNS and SQS
- Design and optimize data storage and querying in Amazon Redshift
- Very good experience in SageMaker and EMR .
- Write performant SQL for data transformations, validation, and reporting
- Ensure data quality, monitoring, error handling and operational support for pipelines
- Collaborate with data consumers, engineers, and stakeholders to translate requirements into solutions
- Contribute to CI / CD , infrastructure-as-code, and documentation for reproducible deployments
Required Skills
Strong experience with Python and Pyspark for large-scale data processingProven hands-on experience with AWS services : Lambda, SNS, SQS, Glue, Redshift, Step FunctionsSolid SQL SQL skills and familiarity with data modeling and query optimizationExperience with ETL best practices, data quality checks, and monitoring / alertingFamiliarity with version control (Git) and basic DevOps / CI-CD workflows