Information Technology
Contract
YASH Technologies
Overview
YASH Technologies is a leading technology integrator specializing in helping clients reimagine operating models, enhance competitiveness, optimize costs, foster exceptional stakeholder experiences, and drive business transformation.
At YASH, we’re a cluster of the brightest stars working with cutting-edge technologies. Our purpose is anchored in a single truth – bringing real positive changes in an increasingly virtual world and it drives us beyond generational gaps and disruptions of the future.
We are looking forward to hire Pentaho Data Integration Professionals in the following areas :
Job Description
Experience required: 4-5 years
Duration: 3 months
Job Summary
The Contractual Data Engineer will play a supportive role in migrating the existing data pipelines from Pentaho Data Integration (PDI/Spoon) to Azure Data Factory (ADF). This individual will primarily assist the core team with data validation, documentation, and the development of new pipelines under guidance.
Key Expectations & Tasks
Our Hyperlearning workplace is grounded upon four principles
At YASH, we’re a cluster of the brightest stars working with cutting-edge technologies. Our purpose is anchored in a single truth – bringing real positive changes in an increasingly virtual world and it drives us beyond generational gaps and disruptions of the future.
We are looking forward to hire Pentaho Data Integration Professionals in the following areas :
Job Description
Experience required: 4-5 years
Duration: 3 months
Job Summary
The Contractual Data Engineer will play a supportive role in migrating the existing data pipelines from Pentaho Data Integration (PDI/Spoon) to Azure Data Factory (ADF). This individual will primarily assist the core team with data validation, documentation, and the development of new pipelines under guidance.
Key Expectations & Tasks
- Understanding Existing Pentaho Pipelines:
- Familiarize with the existing Pentaho Data Integration (PDI/Spoon) jobs and transformations.
- Document the source systems, target systems, data transformations, and business rules implemented in current Pentaho pipelines.
- Identify and document data sources, data types, and data dependencies within the existing Pentaho environment.
- Azure Data Factory (ADF) Development Support:Under guidance, assist in the creation and configuration of ADF pipelines, datasets, linked services, and data flows.
- Translate Pentaho transformation logic into ADF activities (e.g., Copy Data, Data Flow, Stored Procedure, Custom activities).
- Participate in the development of data ingestion, transformation, and loading processes within ADF.
- Execute test cases to validate data integrity and accuracy after migration.
- Compare source data (from Pentaho processed outputs) with target data (from ADF processed outputs) to identify discrepancies.
- Document and report any data discrepancies or issues found during testing
- Assist in troubleshooting and debugging data pipeline issues.
- Contribute to the creation of detailed documentation for migrated ADF pipelines, including design specifications, data lineage, and operational runbooks.
- Document best practices and lessons learned during the migration process.
- Participate in knowledge transfer sessions to the broader team.
- Collaborate effectively with the core development team, data engineers, and other stakeholders.
- Provide regular updates on progress and any roadblocks encountered.
- Strong understanding of ETL/ELT concepts, data warehousing principles, and data modeling.
- Hands-on experience in developing and troubleshooting jobs and transformations in Pentaho.
- Proficient in writing complex SQL queries for data extraction, manipulation, and validation (e.g., joins, subqueries, aggregations).
- Ability to analyze complex data flows, identify issues, and propose solutions.
- Meticulous approach to data validation and documentation.
- Good written and verbal communication skills.
- Familiarity with ADF concepts, activities, and development environment.
- Understanding of cloud platforms (Azure preferred) and data storage services (e.g., Azure Blob Storage, Azure Data Lake Storage).
- Experience with Git or Azure DevOps for code management.
- Basic scripting skills (e.g., Python, PowerShell) for automation or data manipulation.
Our Hyperlearning workplace is grounded upon four principles
- Flexible work arrangements, Free spirit, and emotional positivity
- Agile self-determination, trust, transparency, and open collaboration
- All Support needed for the realization of business goals,
- Stable employment with a great atmosphere and ethical corporate culture
Similar Jobs
View All
Talk to us
Feel free to call, email, or hit us up on our social media accounts.
Email
info@antaltechjobs.in