
Overview
Job Title: PySpark Data Engineer (Remote – India)
Location: Remote (India)
About the Role:
We are looking for a skilled PySpark Data Engineer to join our remote team. This role is ideal for candidates with strong hands-on experience in building and optimizing scalable data pipelines using PySpark and Google Cloud Platform (GCP) tools. You will play a key role in transforming data for advanced analytics and ensuring high performance in data workflows.
Key Responsibilities:
- Design, develop, and maintain ETL/ELT pipelines using PySpark, Dataproc, Cloud Composer (Airflow), and BigQuery
- Optimize and troubleshoot complex SQL queries and data processing workflows
- Implement DevOps best practices for data pipeline deployments
- Work independently and collaborate with cross-functional teams to gather and translate data requirements
- Lead technical discussions and participate in architecture reviews
- (Preferred) Develop event-driven pipelines using Cloud Pub/Sub and Cloud Functions
Required Skills and Experience:
- 7+ years of experience in data engineering
- 4+ years of hands-on experience with GCP tools: BigQuery, Dataproc, Cloud Composer, Cloud Storage
- Strong expertise in PySpark
- Proficiency in optimizing SQL and large-scale data transformation
- Solid understanding of DevOps processes in data engineering environments
Preferred:
- Experience with Cloud Pub/Sub and Cloud Functions
- Prior experience working in offshore/remotely distributed teams
Contact:
Mani - Team Lead HR Recruiter
Email: mani.t@infowaygroup.com
Cell: +91-9513618504
WhatsApp: +91-9513618504
Info Way Solutions LLC
“Email is the best way to reach me”
Job Type: Contractual / Temporary
Contract length: 12 months
Pay: ₹1,500,078.64 - ₹1,841,752.71 per year
Benefits:
- Flexible schedule
- Health insurance
- Paid time off
- Provident Fund
- Work from home
Schedule:
- Day shift
- Monday to Friday
- Morning shift
Education:
- Bachelor's (Preferred)
Experience:
- data engineering: 7 years (Required)
- PySpark: 3 years (Required)
- GCP tools: BigQuery, Dataproc, Cloud Composer, Cloud Storage: 4 years (Required)
- DevOps processes: 4 years (Required)
- SQL: 4 years (Required)
- Cloud Pub/Sub and Cloud Functions: 4 years (Preferred)
- ETL/ELT: 3 years (Required)
Work Location: Remote