Overview
Job Information:
Work Experience: 3+ years
Industry: IT Services
Job Type: FULL TIME
Location: Remote, India
Job Summary:
As a Data Engineer, you’ll manage the invisible backbone of our analytics, building scalable data pipelines that fuel real-time insights for 630+ ships worldwide. Your work will turn raw data points into actionable intelligence, propelling machine learning models and AI-driven decisions that redefine how global operate. This isn’t just ETL; it’s mission-critical infrastructure. You’ll engineer the systems that feed our NOVA analytics dashboards for both our office colleagues and clients — ensuring our PARIS platform stays leagues ahead in a data-driven industry. If you’re the type who thrives on taming chaotic data streams into crystal-clear insights, let’s build the lighthouse that guides maritime innovation.
Key Responsibilities:
Build & Optimise Data Systems
(Focus: Core engineering work with a balance of pipelines and infrastructure)
- Create & maintain data highways: Develop and manage cloud-based data lakes, warehouses, and ETL/ELT pipelines [VK1] that ingest, process, and deliver data from 630+ ships and external sources.
- Keep systems shipshape: Monitor cloud infrastructure performance, resolve bottlenecks, and ensure scalability/reliability for 24/7 maritime operations—no room for “set and forget.”
- Secure the cargo: Implement data quality checks, encryption, and compliance standards (GDPR, SOC2) to protect sensitive maritime telemetry and operational data.
- Automate the mundane: Use tools like Airflow to streamline workflows and reduce manual intervention in pipeline maintenance.
Support Analytics & Troubleshoot Issues
(Focus: Enabling insights while keeping systems running smoothly)
- Fuel AI/ML engines: Partner with Data Scientists to prep datasets for predictive models (e.g., fuel efficiency, preventative maintenance, etc.) and troubleshoot pipeline issues impacting their work.
- Solve data mysteries: Diagnose root causes of pipeline failures, data discrepancies, or MLOps hiccups—then implement fixes that prevent repeat headaches.
- Map the data terrain: Document source-to-target mappings, conduct data profiling, and clarify dependencies so analysts can self-serve without guesswork.
- Stay curious: Experiment with new tools and techniques to improve data quality, system performance and pipeline resilience.
Collaborate & Learn
(Focus: Teamwork and growth in a fast-paced environment)
- Be the glue: Work closely with onshore and offshore developers, IT Operations, and stakeholders across to deliver solutions that balance technical rigor with real-world usability.
- Communicate clearly: Break down complex data concepts for non-technical audiences (think ship captains, not just engineers) and ask questions to avoid ambiguity.
- Learn by doing: Shadow senior engineers, participate in code reviews, and absorb best practices to level up your craft—no prior maritime experience required, but curiosity is a must.
Experience & Qualifications:
- 3+ years’ experience in a Data Engineering role using SQL, PySpark and AirFlow.
- Strong understanding in Data Lake and Data Warehouse design best practices and principles.
- Practical hands-on experience in cloud-based data services for ETL/ELT covering AWS EC2, S3 Storage and EMR.
- Ability to manage and enhance infrastructure related to environments covering Spark, Hive, Presto.
- Experience with databases such as Postgres, MySQL, Oracle.
- Strong work ethic and ability to work independently on agreed goals.
- Clear communication skills in English – both in speaking and writing.
Desirable:
- Deployment experience and management of MLOps framework, such as AWS SageMaker AI, ECR.
- Experience in other cloud platform and hybrid cloud infrastructure, e.g. GCP, Azure.
- Experience in the maritime industry.
Competencies:
- Analysis & Problem Solving (Level 2): Uses critical thinking to address problems. Able to perform root cause analysis on complex problems to identify underlying trends and put forward well-thought out solutions to address the causes.
- Listening & Communication (Level 2): Focuses on the individual they are communicating with. Writes and expresses thoughts clearly adjusting as necessary to the audience. Asks questions to clarify.
- Collaboration, Inclusion & Teamwork (Level 1): A good team player that is personable, friendly, polite and takes the time to know people. Collaborates with onshore and offshore colleagues well.
- Customer Focus (Level 2): Understands the needs of the customer clarifying requirements and expectations. Adapts as necessary to changing requirements and is responsive, helpful with all requests. Sets high quality for service delivery.
- Planning & Organising (Level 2): Uses the supplied tools for structured project planning for optimal time use. Balances competing priorities. Promptly Updates people when plans change and keeps them apprised of progress. Adjusts own plans based on FML strategies and plans.
- Initiative (Level 2): Challenges existing ways of doing things. Looking for continuous improvement without being asked. Identifies ways to improve efficiency and effectiveness in own work and others in the team.
- Accountability (Level 2): Responsible for delivery own work without unnecessary supervision. Always demonstrates integrity and ethical behavior being a role model for others. Invites constructive feedback adjusting actions accordingly. Thinks through mistakes made and finds ways to prevent reoccurrence.
Interview Process
- 2 Assessments
- 3 Technical Rounds