Free cookie consent management tool by TermsFeed Databricks Data Engineer | Antal Tech Jobs
Back to Jobs
1 Day ago

Databricks Data Engineer

decor
Pune, Maharashtra, India
Information Technology
Full-Time
Astellas Pharma

Overview

As part of the Astellas commitment to delivering value for our patients, our organisation is currently undergoing transformation to achieve this critical goal. This is an opportunity to work on digital transformation and make a real impact within a company dedicated to improving lives.

DigitalX our new information technology function is spearheading this value driven transformation across Astellas. We are looking for people who excel in embracing change, manage technical challenges and have exceptional communication skills.

We are seeking committed and talented DataBricks Developers with 2- 4 years progressive experience to join our new InformationX team- which lies at the heart of DigitalX.

The ideal candidate will have a minimum of 2 years of professional experience leveraging the Databricks platform to deliver data engineering, machine learning (ML), and business intelligence outputs. You will be responsible for building and maintaining robust data pipelines, developing scalable ML models, and generating actionable insights from large datasets. This role requires a strong understanding of big data technologies, data architecture, and proficiency in languages such as Python and SQL.

As a member of our team within InformationX, you will be responsible for ensuring our data driven systems are operational, scalable and continue to contain the right data to drive business value.

Responsibilities

Your responsibilities will include executing complex data projects, ensuring smooth data flows between systems, and maintaining the efficiency and reliability of data platforms. This is a fantastic global opportunity to use your proven agile delivery skills across a diverse range of initiatives, utilise your development skills, and contribute to the continuous improvement/delivery of critical IT (Information Technology) solutions.

  • Data Engineering: Design, develop, and maintain efficient and reliable ETL/ELT pipelines using Databricks notebooks and Delta Lake.
  • Machine Learning: Collaborate with data scientists to deploy and manage scalable ML models, ensuring they are integrated into production workflows.
  • Insight Delivery: Create and optimize notebooks and queries to provide data-driven insights and reports to business stakeholders.
  • Platform Management: Manage Databricks clusters and jobs, ensuring optimal performance, cost-efficiency, and security.
  • Collaboration: Work closely with data scientists, analysts, and business teams to understand requirements and deliver end-to-end data solutions.

You Will Also Be Contributing To The Following Areas

  • End-to-End Data Solutions:
  • Design end-to-end scalable data streams, storage, data serving systems, and analytical workflows using Databricks.
  • Define overall architecture, capabilities, platforms, tools, and governing processes.
  • Data Pipeline Development:
  • Build data pipelines to extract, transform, and load data from various sources.
  • Set up metadata and master data structures to support transformation pipelines in Databricks.
  • Data Warehousing and Data Lakes:
  • Create data warehouses and data lakes for efficient data storage and management.
  • Develop and deploy data processing and analytics tools.
  • Collaboration with DataX and other key stakeholder value teams:
  • Collaborate with data modelers to create advanced data structures and models within the Databricks environment.
  • Develop and maintain Python scripts for data processing, transformation, and analysis.
  • Utilize Azure and AWS cloud services (e.g., Azure Data Lake, AWS S3, Redshift) for data storage and processing.
  • Apply expertise in Databricks to enhance data architecture, performance, and reliability. Lead relevant data governance initiatives and ensure compliance with industry standards.
  • Work closely with data scientists to develop and deploy data-driven solutions.
  • Provide technical direction to Data Engineers and perform code reviews.
  • Continuous Learning:
  • Stay up to date on the latest data technologies, trends, and best practices.
  • Participate in smaller focused mission teams to deliver value driven solutions aligned to our global and bold move priority initiatives and beyond.
  • Collaborate with cross functional teams and practises across the organisation including Commercial, Manufacturing, Medical, DataX, GrowthX and support other X (transformation) Hubs and Practices as appropriate, to understand user needs and translate them into technical solutions.
  • Provide Level 3 and 4 Technical Support to internal users troubleshooting complex issues and ensuring system uptime as soon as possible.
  • Champion continuous improvement initiatives identifying opportunities to optimise performance security and maintainability of existing data and platform architecture and other technology investments.
  • Participate in the continuous delivery pipeline Adhering to DevOps best practises for version control automation and deployment. Ensuring effective management of the FoundationX backlog.
  • Leverage your knowledge of Machine Learning (ML) and data engineering principles to integrate with existing data pipelines and explore new possibilities for data utilization.
  • Stay up to date on the latest trends and technologies in full-stack-development, data engineering and cloud platforms.

Technical Skills

  • Proficiency in PySpark/ Python or Scala for data manipulation, scripting, and analytics
  • Strong understanding of distributed computing principles
  • Experience using ETL tools like Talend/Talend Cloud and DataStage.
  • Knowledge and experience using Azure DevOps.
  • Experience with cloud services such as Azure Data Lake, AWS S3, and Redshift
  • Experience in working with MPP Databases like AWS Redshift.
  • Experience in integrating data from multiple sources like relational databases, Salesforce, SAP, and API calls.

Required Qualifications

  • Bachelor's or Master's degree in Computer Science, Engineering, or related field.
  • 3 years+ of experience as a Data Engineer or DataBricks Developer.
  • Proficiency in Python for data manipulation, scripting, and analytics.
  • Strong understanding of data modelling concepts and practices.

Any relevant cloud-based DataBricks, AWS or Azure certifications for example:

  • Databricks Data Engineer
  • AWS Certified Data Analytics Speciality – Professional / Associate (will be considered with relevant experience)
  • Microsoft Certified Azure Data Engineer Associate
  • Microsoft Certified Azure Database Administrator
  • Microsoft Certified Azure Developer
  • Experience using ETL tools like Talend / Talend Cloud and DataStage (Essential)
  • Knowledge and experience using Azure DevOps (Essential)
  • Knowledge and experience of working with SalesForce / SAP (Desirable)
  • Experience in working with MPP Databases like AWS Redshift
  • Experience of delivering architectural solutions effectively within Lifesciences or Pharma Domains.

Preferred Qualifications

  • Experience analysing and building star schema data warehouses
  • Experiencing writing SQL and creating stored procedures is essentialData Analysis and Automation Skills: Proficient in identifying, standardizing, and automating critical reporting metrics and modelling tools.
  • Analytical Thinking: Demonstrated ability to lead ad hoc analyses, identify performance gaps, and foster a culture of continuous improvement.
  • Experience in integrating data from multiple Data sources like relational Databases, Salesforce, SAP and API calls.
  • Agile Champion: Adherence to DevOps principles and a proven track record with CI/CD pipelines for continuous delivery.
  • Understand and Interpret business requirements and can term them into technical requirements.
  • Create and maintain technical documentation as part of CI/CD principles

"Beware of recruitment scams impersonating Astellas recruiters or representatives. Authentic communication will only originate from an official Astellas LinkedIn profile or a verified company email address. If you encounter a fake profile or anything suspicious, report it promptly to LinkedIn's support team through LinkedIn Help"

Category

Astellas is committed to equality of opportunity in all aspects of employment.

EOE including Disability/Protected Veterans

Share job
Similar Jobs
View All
1 Day ago
Senior Network Engineer
Information Technology
  • Pune, Maharashtra, India
We are looking for a capable Network Engineer, who enjoys security work and possesses both deep and wide expertise in the security space.Qualified candidates will have a background in Network Devices handling and Engineering. Roles & Responsibilitie...
decor
1 Day ago
Cloud Support Engineer EMEA
Information Technology
  • Pune, Maharashtra, India
OverviewThe Cloud Support Engineer - EMEA isn’t just an ordinary Support Engineer role. As a Support Engineer you will provide multi-channel support, including via chat and email, to customers using Atlassian products. Your responsibilities will incl...
decor
1 Day ago
SQL Developer (Senior)
Information Technology
  • Pune, Maharashtra, India
Roles & ResponsibilitiesThe SQL developer senior is a highly experienced SQL programmer who can design and implement complex database structures, write optimized SQL queries, and lead projects related to databases.Experience 8-11 YearsSkillsPrimary ...
decor
1 Day ago
Custom Software Engineer
Information Technology
  • Pune, Maharashtra, India
Project Role : Custom Software EngineerProject Role Description : Design, build and configure applications to meet business process and application requirements.Must have skills : SAP ABAP Development for HANAGood to have skills : NAMinimum 7.5 Year(...
decor
1 Day ago
Senior Software Engineer (PostgreSQL, Oracle EBS)
Information Technology
  • Pune, Maharashtra, India
Job DescriptionExpect more. Connect more. Be more at Diebold Nixdorf. Our teams automate, digitize, and transform the way more than 75 million people around the globe bank and shop in this hyper-connected, consumer-centric world. Join us in connectin...
decor
1 Day ago
Data Engineer
Information Technology
  • Pune, Maharashtra, India
Who We AreAt Kyndryl, we design, build, manage and modernize the mission-critical technology systems that the world depends on every day. So why work at Kyndryl? We are always moving forward – always pushing ourselves to go further in our efforts to ...
decor
1 Day ago
Python Data Curation Engineer
Information Technology
  • Pune, Maharashtra, India
Role OverviewDesigns, builds and oversees the deployment and operation of technology architecture, solutions and software to capture, manage, store and utilize structured and unstructured data from internal and external sources. Reviews internal and ...
decor
1 Day ago
Lead Data Scientist
Information Technology
  • Pune, Maharashtra, India
Company: IT Services OrganizationKey Skills: Data Science & ML: TensorFlow, Spark ML, H20, Machine Learning, Deep Learning, Gen AI, LLMs, MLOps, Data Scientist, Data modeling, Data Science, Data Scientist R Programming, AWS, GCP, Cloud (AWS / Azure /...
decor

Talk to us

Feel free to call, email, or hit us up on our social media accounts.
Social media