1000000 - 1000000 Indian Rupee - Yearly
Indore, Madhya Pradesh, India
Finance & Banking
Full-Time
Go Digital Technology Consulting LLP
Overview
Role: Senior Data Engineer
Location: Mumbai
Experience: 4yrsto 8yrs
Technologies / Skills: Advanced SQL, Python and associated librarieslike Pandas, Numpy etc., Pyspark , Shell scripting, DataModelling, Big data, Hadoop, Hive, ETL pipelines.
Responsibilities:
- Proven successin communicating with users, other technical teams, and senior management to collect requirements, describe data modeling decisions and develop data engineering strategy.
- Ability to work with business ownersto define key businessrequirements and convert to user stories with required technical specifications.
- Communicate results and businessimpacts of insight initiatives to key stakeholders to collaboratively solve business problems.
- Working closely with the overall Enterprise Data & Analytics Architect and Engineering practice leads to ensure adherence with the best practices and design principles.
- Assures quality, security and compliance requirements are met forsupported area.
- Design and create fault-tolerance data pipelinesrunning on cluster
- Excellent communication skills with the ability to influence client business and IT teams
- Should have design data engineering solutions end to end. Ability to come up with scalable and modular solutions
Required Qualification:
- 3+ years of hands-on experience Designing and developing Data Pipelinesfor Data Ingestion or Transformation using Python (PySpark)/Spark SQL in AWS cloud
- Experience in design and development of data pipelines and processing of data at scale.
- Advanced experience in writing and optimizing efficient SQL queries with Python and Hive handling Large Data Sets in Big-Data Environments
- Experience in debugging, tunning and optimizing PySpark data pipelines
- Should have implemented concepts and have good knowledge of Pyspark data frames, joins, caching, memory management, partitioning, parallelism etc.
- Understanding of Spark UI, Event Timelines, DAG, Spark config parameters, in order to tune the long running data pipelines.
- Experience working in Agile implementations
- Experience with building data pipelinesin streaming and batch mode.
- Experience with Git and CI/CD pipelines to deploy cloud applications
- Good knowledge of designing Hive tables with partitioning for performance.
Desired Qualification:
- Experience in data modelling
- Hands on creating workflows on any Scheduling Tool like Autosys, CA Workload Automation
- Proficiency in using SDKsfor interacting with native AWS services
- Strong understanding of concepts of ETL, ELT and data modeling.
Job Types: Full-time, Permanent
Pay: From ₹1,000,000.00 per year
Benefits:
- Commuter assistance
- Health insurance
- Internet reimbursement
- Paid sick time
- Paid time off
- Provident Fund
Schedule:
- Monday to Friday
Work Location: In person
Similar Jobs
View All
Talk to us
Feel free to call, email, or hit us up on our social media accounts.
Email
info@antaltechjobs.in