
Overview
Data Engineer / FS - Engineer (PySpark | Databricks | ADF | Big Data | Hadoop | Hive)
Location : Hyderabad , Noida, Gurgaon, Bangalore.
Job Type : Contractual Hybrid
Experience Required: 8+ Years
Shift Timing : 11 am – 8:30 pm
Job Description:
We are looking for a skilled Data Engineer / Full Stack Data Engineer with strong experience in PySpark, Databricks, Azure Data Factory (ADF), Big Data Ecosystems, Hadoop, and Hive to join our growing data engineering team. The ideal candidate will have a strong grasp of distributed data processing and cloud data platforms, especially within the Azure ecosystem.
Key Responsibilities:
- Design, build, and maintain scalable data pipelines using PySpark and Databricks.
- Develop and orchestrate ETL workflows using Azure Data Factory (ADF).
- Work on Big Data technologies including Hadoop and Hive for large-scale data processing.
- Ensure performance optimization and tuning of data processes and jobs.
- Collaborate with data scientists, analysts, and business stakeholders to deliver reliable data solutions.
- Implement best practices in data engineering, including data quality, logging, monitoring, and alerting.
- Troubleshoot and debug performance issues in data pipelines and environments.
- Manage code in version control systems like Git and follow CI/CD practices.
Required Skills & Experience:
- Strong hands-on experience with PySpark and Databricks.
- Proficient in Azure Data Factory (ADF) and Azure cloud ecosystem.
- In-depth knowledge of Big Data technologies like Hadoop, Hive.
- Experience with data modeling, data warehousing, and building large-scale data systems.
- Strong understanding of data governance, data quality, and security best practices.
- Good knowledge of SQL and performance tuning in large data sets.
- Familiarity with Git, DevOps pipelines, and agile methodologies.
Preferred Qualifications:
- Azure or Databricks certifications.
- Experience with other Azure services such as ADLS, Azure Synapse, Azure Functions.
- Exposure to scripting in Python beyond PySpark (for automation or orchestration).
- Understanding of Delta Lake, Data Lake architecture, and partitioning strategies.
Note: The candidate must go to the client location to collect the laptop.
Job Types: Full-time, Contractual / Temporary
Contract length: 6 months
Pay: ₹85,000.00 - ₹100,000.00 per month
Schedule:
- Day shift
Work Location: In person