Design and maintain a unified data architecture: database schemas, data models, and micro-architecture solutions to ensure scalability and reliability.
Optimize database performance at all levels: indexing, partitioning, clustering, and tuning configuration parameters.
Ensure full compliance with GDPR, UK Data Protection Act, and other relevant regulations: data masking, consent management, retention policies, and privacy impact assessments
Optimize queries, schemas, and indexes where needed
Set up basic data quality checks
Support GDPR and UK data protection requirements, including: Data masking, Access control, Retention policies
Take data notebooks and calculation logic
Turn them into reliable, production-ready pipelines
Ensure scalability, reliability, and reproducibility
Requirements
Write clean, readable, maintainable code
Have real experience supporting data pipelines in production
Have worked with a data warehouse (BigQuery or similar)
Have strong experience in GCP
Understand orchestration, monitoring, and performance tuning
Can make practical engineering decisions independently
Strong communication skills and fluency in English.
Startup mentality: resilience, adaptability, and ability to thrive in a fast-paced environment.
Customer-centric mindset: focus on delivering value to end-users or clients.
Strong problem-solving skills – ability to approach challenges logically and propose practical solutions.
Nice to have: Experience with AWS, or Azure
Nice to have: Experience with message queues or distributed systems
Nice to have: Basic CI/CD for data pipelines
Tech Stack
AWS
Azure
BigQuery
Distributed Systems
Google Cloud Platform
Benefits
The role is fully remote, providing flexibility and enabling seamless collaboration with our geographically distributed team.
Competitive salary with the potential for equity options based on performance, recognising exceptional contributions to our integration success.