Overview
Purpose of the role
To build and maintain the systems that collect, store, process, and analyze data, such as data pipelines, data warehouses and data lakes to ensure that all data is accurate, accessible, and secure.
Accountabilities
- Build and maintenance of data architectures pipelines that enable the transfer and processing of durable, complete and consistent data.
- Design and implementation of data warehoused and data lakes that manage the appropriate data volumes and velocity and adhere to the required security measures.
- Development of processing and analysis algorithms fit for the intended data complexity and volumes.
- Collaboration with data scientist to build and deploy machine learning models.
Assistant Vice President Expectations
- To advise and influence decision making, contribute to policy development and take responsibility for operational effectiveness. Collaborate closely with other functions/ business divisions.
- Lead a team performing complex tasks, using well developed professional knowledge and skills to deliver on work that impacts the whole business function. Set objectives and coach employees in pursuit of those objectives, appraisal of performance relative to objectives and determination of reward outcomes
- Collaborate with other areas of work, for business aligned support areas to keep up to speed with business activity and the business strategy.
- Engage in complex analysis of data from multiple sources of information, internal and external sources such as procedures and practises (in other areas, teams, companies, etc).to solve problems creatively and effectively.
- Communicate complex information. 'Complex' information could include sensitive information or information that is difficult to communicate because of its content or its audience.
- Influence or convince stakeholders to achieve outcomes.
As data Engineer. role focuses on building scalable, secure, and high‑performance data pipelines that power enterprise‑grade Generative AI solutions. You will work across Python, PySpark, AWS, Glue, and modern data engineering tooling, enabling rapid experimentation, safe deployment, and high-quality data for AI/ML workloads. While also building reporting, insights, and analytics layers that help measure performance, adoption, and effectiveness of GenAI systems.
To be successful in this role you should have:
- Strong experience with Python (Expert) , PySpark, and AWS Glue.
- Solid understanding of distributed data processing, Spark optimization, and scalable data design.
- Hands-on AWS data engineering experience (S3, EMR, Lambda, Step Functions, Athena).
- Proficiency in SQL for analytical modelling and reporting dataset creation.
- Experience building dashboards or following requirements from BI teams.
- Understanding of vectorization, RAG pipelines, embeddings, and basic LLM concepts.
- Familiarity with CI/CD, Git, code quality, and automated testing for data pipelines.
- Experience with Kafka/Kinesis for streaming analytics.
Some other highly valued skills include:
- Strong communication and stakeholder management skills.
- Ability to lead technical teams and mentor junior developers.
- Collaborate with GenAI engineers, platform teams, ML engineers, data scientists, and product managers.
- Bachelor’s degree in computer science, Information Technology or Engineering is required.