Bangalore, Karnataka, India
Information Technology
Full-Time
ADM
Overview
Job Description
Your Responsibilities
We are seeking a skilled and motivated Azure Databricks Sr Data Engineer to join our dynamic team. The ideal candidate will have strong experience with Python, Spark programming, and expertise in building and optimizing data pipelines in Azure Databricks. You will play a pivotal role in leveraging Databricks workflows, Databricks Asset Bundles, and CI/CD pipelines using GitHub to deliver high-performance data solutions. A solid understanding of Data Warehousing and Data Mart architecture in Databricks is critical for success in this role. If you’re passionate about data engineering, cloud technologies, and scalable data architecture, we’d love to hear from you!
Diversity, equity, inclusion and belonging are cornerstones of ADM’s efforts to continue innovating, driving growth, and delivering outstanding performance. We are committed to attracting and retaining a diverse workforce and create welcoming, truly inclusive work environments — environments that enable every ADM colleague to feel comfortable on the job, make meaningful contributions to our success, and grow their career. We respect and value the unique backgrounds and experiences that each person can bring to ADM because we know that diversity of perspectives makes us better, together.
For more information regarding our efforts to advance Diversity, Equity, Inclusion & Belonging, please visit our website here: Diversity, Equity and Inclusion | ADM.
About ADM
At ADM, we unlock the power of nature to provide access to nutrition worldwide. With industry-advancing innovations, a complete portfolio of ingredients and solutions to meet any taste, and a commitment to sustainability, we give customers an edge in solving the nutritional challenges of today and tomorrow. We’re a global leader in human and animal nutrition and the world’s premier agricultural origination and processing company. Our breadth, depth, insights, facilities and logistical expertise give us unparalleled capabilities to meet needs for food, beverages, health and wellness, and more. From the seed of the idea to the outcome of the solution, we enrich the quality of life the world over. Learn more at www.adm.com.
Req/Job ID
97476BR
Ref ID
Your Responsibilities
We are seeking a skilled and motivated Azure Databricks Sr Data Engineer to join our dynamic team. The ideal candidate will have strong experience with Python, Spark programming, and expertise in building and optimizing data pipelines in Azure Databricks. You will play a pivotal role in leveraging Databricks workflows, Databricks Asset Bundles, and CI/CD pipelines using GitHub to deliver high-performance data solutions. A solid understanding of Data Warehousing and Data Mart architecture in Databricks is critical for success in this role. If you’re passionate about data engineering, cloud technologies, and scalable data architecture, we’d love to hear from you!
- Python and Spark Programming:
- Develop and maintain scalable data pipelines using Python and Apache Spark within Azure Databricks.
- Write optimized, high-performance Spark jobs to process large volumes of data efficiently.
- Utilize PySpark for distributed data processing, transformation, and aggregation tasks.
- Databricks Workflows:
- Design and implement Databricks Workflows to automate data pipeline execution, orchestrating complex workflows and batch jobs.
- Set up task dependencies, triggers, and notifications to ensure smooth and reliable execution.
- Monitor, troubleshoot, and optimize Databricks workflows for optimal performance and minimal failures.
- Databricks Asset Bundles:
- Create and manage reusable components such as Databricks Asset Bundles, including notebooks, libraries, and models.
- Share and reuse asset bundles across teams to increase efficiency and ensure consistency in development.
- CI/CD for Databricks Artifacts using GitHub:
- Implement CI/CD pipelines using GitHub Actions for the continuous integration and deployment of Databricks notebooks, jobs, and libraries.
- Automate the testing, building, and deployment processes to ensure smooth, consistent code delivery across environments.
- Collaborate with teams to implement version control practices and code reviews using GitHub.
- Data Warehousing & Data Mart Design:
- Design and implement Data Warehousing and Data Mart solutions using Databricks, ensuring high-performance storage and retrieval of structured data.
- Integrate data from multiple sources into a central data warehouse using Spark-based transformations, ensuring efficient schema design and query performance.
- Implement dimensional modeling, including star and snowflake schemas, within Azure Databricks for data marts to support business intelligence and reporting.
- Data Pipeline Optimization and Management:
- Continuously monitor and optimize Databricks-based data pipelines for performance, scalability, and cost efficiency.
- Implement best practices for data partitioning, caching, and query optimization within the Databricks platform.
- Troubleshoot and resolve issues related to data integrity, performance, and workflow execution.
- Collaboration and Stakeholder Communication:
- Work closely with data scientists, analysts, and other teams to understand requirements and build data solutions that meet business needs.
- Communicate technical concepts effectively to both technical and non-technical stakeholders.
- Provide mentorship and guidance to junior data engineers on Databricks best practices, data architecture, and efficient coding techniques.
- Python and Spark Programming:
- Minimum of 3 years of experience in Python programming, especially in data engineering, ETL processes, and distributed computing.
- Solid experience using Apache Spark (PySpark) for large-scale data processing and transformation within Databricks.
- Proficiency in writing and optimizing Spark-based jobs for high performance on large datasets.
- Databricks Workflows:
- Strong hands-on experience with Databricks Workflows for orchestrating data pipelines and batch processes.
- Ability to design and optimize multi-step workflows with task dependencies, retries, and monitoring.
- Databricks Asset Bundles:
- Experience in creating and managing Databricks Asset Bundles to promote reusability and modularization of notebooks, libraries, and models.
- CI/CD for Databricks Artifacts using GitHub:
- Experience with implementing CI/CD pipelines for Databricks using GitHub and GitHub Actions for automating deployment of notebooks, jobs, and libraries.
- Expertise in version control practices and integrating Databricks with external Git repositories for collaborative development.
- Data Warehousing & Data Mart Experience:
- Strong experience in designing and implementing Data Warehouses and Data Marts using Databricks and Spark.
- Understanding of dimensional modeling (star and snowflake schemas) and the ability to create optimized data structures for reporting and analytics.
- Hands-on experience integrating data from multiple sources and managing the ETL process within a Data Warehouse or Data Mart environment.
- Cloud Experience:
- Solid experience working with the Azure ecosystem, including Azure Data Lake, Azure Blob Storage, and Azure SQL Database.
- Experience working in cloud environments and leveraging cloud-based tools for building and managing data pipelines.
- Data Engineering Best Practices:
- Knowledge of best practices for designing and managing scalable, efficient, and cost-effective data pipelines.
- Experience in performance tuning and query optimization within Databricks and Spark.
- Collaboration and Communication:
- Excellent teamwork and communication skills, with the ability to collaborate effectively across cross-functional teams.
- Ability to document technical processes and communicate progress and results to stakeholders.
- Cloud Certifications:
- Azure certifications, particularly in Databricks, Data Engineering, or Cloud Solutions, are a plus.
- Big Data Technologies:
- Familiarity with other big data tools such as Kafka, Hadoop, or Flink for streaming and real-time data processing is a plus.
- Data Science/ML Experience:
- Exposure to machine learning workflows and model management within Databricks (e.g., using MLflow) is beneficial.
Diversity, equity, inclusion and belonging are cornerstones of ADM’s efforts to continue innovating, driving growth, and delivering outstanding performance. We are committed to attracting and retaining a diverse workforce and create welcoming, truly inclusive work environments — environments that enable every ADM colleague to feel comfortable on the job, make meaningful contributions to our success, and grow their career. We respect and value the unique backgrounds and experiences that each person can bring to ADM because we know that diversity of perspectives makes us better, together.
For more information regarding our efforts to advance Diversity, Equity, Inclusion & Belonging, please visit our website here: Diversity, Equity and Inclusion | ADM.
About ADM
At ADM, we unlock the power of nature to provide access to nutrition worldwide. With industry-advancing innovations, a complete portfolio of ingredients and solutions to meet any taste, and a commitment to sustainability, we give customers an edge in solving the nutritional challenges of today and tomorrow. We’re a global leader in human and animal nutrition and the world’s premier agricultural origination and processing company. Our breadth, depth, insights, facilities and logistical expertise give us unparalleled capabilities to meet needs for food, beverages, health and wellness, and more. From the seed of the idea to the outcome of the solution, we enrich the quality of life the world over. Learn more at www.adm.com.
Req/Job ID
97476BR
Ref ID
Similar Jobs
View All
Talk to us
Feel free to call, email, or hit us up on our social media accounts.
Email
info@antaltechjobs.in