Overview
Job descriptions may display in multiple languages  based on your language selection.
What we offer:
Group Summary:
Job Responsibilities:
Key Responsibilities
• Design and execute comprehensive test plans for real-time and batch data pipelines handling IoT, robotics, and operational telemetry.
• Validate data ingestion from Kafka and Debezium streams into MongoDB and Databricks Delta Lake .
• Perform data integrity checks on historical and live datasets related to forklift movement, robot actions, and plant sensor events.
• Develop automated data validation scripts using SQL , Python , or PyTest for regression and integration testing.
• Ensure consistency across visualizations in Grafana , QuickSight , and operational dashboards.
• Identify, document, and track data anomalies, schema mismatches, or pipeline failures.
• Validate schema changes across NoSQL and structured databases, ensuring compatibility with downstream systems.
• Work closely with Data Engineers, Site Reliability Engineers, and Manufacturing Ops to maintain high data quality standards.
• Support root cause analysis for incorrect metrics or delayed data pipelines in production.
Required Skills and Experience
• Bachelor’s degree in Computer Science, Data Engineering, Industrial Engineering, or related field.
• 2+ years of experience in data validation, QA, or analytics testing in a manufacturing or IoT environment .
• Strong skills in SQL and Python for data querying, transformation, and testing.
• Experience with data platforms like MongoDB and Databricks (Delta Tables, Notebooks).
• Familiarity with Kafka , Debezium , or other real-time data ingestion tools.
• Exposure to monitoring and visualization tools such as Grafana and Amazon QuickSight .
• Understanding of manufacturing telemetry , machine-to-cloud data flow , and industrial IoT standards.
Preferred Qualifications
• Experience with data test automation frameworks (e.g., PyTest, Great Expectations).
• Familiarity with CI/CD pipelines and version-controlled test environments.
• Knowledge of time-series databases and streaming analytics.
• Understanding of data governance , data contracts , or schema registry practices.
Â