Overview
About the Role
We are looking for a seasoned Data Architect to lead the data architecture track for a large-scale analytics transformation programme. This is a hands-on leadership role — you will be expected to design and build while simultaneously guiding a team, aligning with stakeholders, and driving delivery against a complex multi-domain roadmap.
You will work at the intersection of data engineering, programme governance, and business intelligence. The ideal candidate has strong architectural instincts, deep hands-on technical ability, and the communication skills to operate confidently with both engineering teams and senior business stakeholders.
What You Will Do
Architecture & Design
Own the end-to-end data architecture across multiple layers of a medallion architecture on AWS
Design atomic star schema models — fact and dimension tables — across multiple business domains
Define aggregation and KPI layers aligned to business reporting requirements and industry frameworks
Establish data modelling standards, naming conventions, and design patterns for the team to follow
Ensure architecture decisions align with the broader enterprise data strategy and roadmap
Hands-on Delivery
Write and review SQL, Python, and Spark code for data pipelines and transformations
Perform data reconciliation and quality validation across layers
Build and review data models in Snowflake or Databricks
Work within the AWS ecosystem (Athena, Glue, S3) to deliver scalable, production-grade pipelines
Leadership & Stakeholder Management
Lead and unblock a team of data engineers and analysts on day-to-day delivery
Proactively identify blockers, dependencies, and risks — raise them early with proposed solutions, not just flags
Coordinate across business domains to gather requirements and validate data outputs
Translate business requirements into technical designs and communicate technical decisions back to non-technical stakeholders clearly
Align delivery milestones with the programme roadmap and surface deviations proactively
Maintain architecture documentation in Confluence and manage delivery tasks in Jira
What You Bring
Domain Knowledge
Ability to quickly understand a new business domain, ask the right questions, and translate operational processes into data models
Prior exposure to any transaction-heavy business domain is valuable — examples include Source-to-Pay, Supply Chain, Finance, HR, or Customer Data
Experience working with ERP or enterprise application data (Oracle, SAP, Salesforce, or similar) as a source is an advantage
Familiarity with industry benchmarking or process frameworks (such as APQC or SCOR) is a plus but not required
Technical Skills (Essential)
Advanced SQL — complex joins, window functions, performance tuning
Python — data transformation, automation, pipeline scripting
Apache Spark — large-scale distributed data processing
AWS — Athena, Glue, S3; experience designing cloud-native data platforms
Snowflake or Databricks — data modelling, query optimisation, pipeline orchestration
Data modelling — star schema, dimensional modelling, normalisation, slowly changing dimensions
Data reconciliation — validating data integrity across source and target layers
Full SDLC — requirements, design, build, test, release, operate
Tools
Jira — managing delivery, tracking progress, raising and escalating blockers
Confluence — authoring and maintaining architecture and design documentation
Good to Have
Snowflake or Databricks certification
Exposure to BI tools (Power BI, Tableau) from an architecture or data modelling perspective
The Kind of Person We Are Looking For
You see blockers coming before they arrive and raise them with a solution, not just a flag
You can hold a room — whether that is a sprint planning with engineers or a roadmap review with a VP
You write documentation that people actually read
You are as comfortable doing the work yourself as reviewing someone else's
You care about data quality as much as data delivery
You ask "why does the business need this?" before deciding how to build it