Overview
About the Role
We are looking for an experienced Senior Data Engineer with strong expertise in building and managing modern AWS-based Data Lakehouse architectures. The ideal candidate will have hands-on experience working with large-scale data pipelines, ETL frameworks, and AWS-native data engineering services.
You will play a key role in designing, developing, and optimizing scalable data solutions that support analytics, reporting, and business intelligence use cases.
Key Responsibilities
• Design and implement robust Data Lakehouse solutions on AWS
• Develop scalable ETL pipelines using AWS Glue and PySpark
• Build and maintain efficient data ingestion and transformation workflows
• Work extensively with AWS S3 for data storage and lakehouse structuring
• Perform data preparation and cleansing using AWS Glue DataBrew
• Implement and manage data quality checks using AWS Data Quality tools
• Optimize performance of data pipelines for high-volume processing
• Collaborate with cross-functional teams including analysts, architects, and stakeholders
• Ensure best practices in data governance, security, and compliance
**** Required Skills & Qualifications****
• 7+ years of experience in Data Engineering
• Strong expertise in AWS Data Lakehouse architecture
• Proficiency in Python and PySpark
• Hands-on experience with:
• AWS S3
• AWS Glue ETL
• AWS Glue DataBrew
• AWS Data Quality Frameworks
• Experience in building large-scale batch and/or streaming pipelines
• Strong understanding of data modeling, warehousing, and lakehouse concepts
• Excellent problem-solving and communication skills
Preferred Qualifications:-
• Experience working in Agile development environments
• Exposure to Delta Lake, Apache Iceberg, or similar lakehouse technologies
• AWS Certification (Data Analytics or Solutions Architect) is a plus