Senior Data Platform Engineer
Job Description
Hi, we are Storyteq!
Storyteq helps marketing and creative teams in top brands to scale their creative production and gain control over their marketing campaigns. Through our platform, our clients can streamline campaign workflows, automate their creative production through templates & AI and activate engaging campaigns that go to market faster.
We believe creativity takes time, but creative production shouldn't. Since 2016 our mission has been to revolutionize the way creative assets are created and used. Magic happens when we let technology empower creativity. Our endless curiosity and relentless commitment to our customers lies at the heart of our problem solving approach. This shared mission is woven into our values: we dream big, think differently, and are stronger together.
Role OverviewThe Senior Data Platform Engineer is a hands on technical specialist responsible for implementing and maintaining the data infrastructure that underpins our global platform. Working across a range of database technologies, from relational databases through to vector databases, graph databases, and cloud scale data lakes, this engineer builds reliable, high quality data systems that meet the needs of our product and engineering teams.
A key part of this role is contributing to the build out of an enterprise scale data layer that provides rich, reliable context to our agentic AI product capabilities. As we advance our AI powered platform, the quality, structure, and accessibility of our underlying data is critical, and this engineer will play an important hands on role in ensuring that foundation is sound.
You will work closely with product and engineering teams to understand data requirements, contribute to technology decisions, and implement solutions that are well configured, clearly documented, and built to last. This is a role for an engineer who takes real pride in the craft of data engineering and who cares about the reliability and integrity of the systems they build.
1. Database Implementation & Technology Selection- Work with product and engineering teams to understand data requirements and contribute to technology selection, covering relational databases, vector databases, graph databases, and cloud data lake technologies.
- Implement chosen database technologies, including configuration, schema design, and integration with application and data layers.
- Apply data modelling best practices to ensure schemas are well structured, performant, and maintainable.
- Ensure configurations are optimised for performance, reliability, and cost efficiency, and keep documentation of schemas and architectural decisions up to date.
- Implement validation, constraint, and reconciliation mechanisms to prevent and detect data corruption or inconsistency across distributed data stores.
- Apply established best practices for data integrity and consistency, including idempotency and conflict resolution patterns in distributed systems.
- Monitor data health and address integrity issues promptly before they impact product or customer outcomes.
- Work with application engineers to ensure data quality considerations are factored into how services write and consume data.
- Build and maintain scalable data lake architectures on Google Cloud Platform including data modelling, partitioning, clustering, and cost optimisation.
- Develop robust ETL/ELT pipelines that ingest, transform, and serve data reliably at scale.
- Instrument pipelines with appropriate observability, error handling, retry logic, and lineage tracking.
- Ensure data lake structures are well organised and accessible to support analytics, data science, and AI product use cases.
- Build and maintain automated data quality checks, monitoring, and alerting across pipelines and data stores.
- Work with data consumers, including data science, analytics, and product teams, to understand quality requirements and address issues as they arise.
- Apply data governance practices in day to day work, including appropriate access controls, retention handling, and lineage tracking.
- Ensure sensitive data is handled correctly and in line with security and compliance requirements.
- Solid hands on experience with relational database technologies including schema design, query optimisation, indexing, and day to day operational management.
- Practical experience with vector databases and vector search technologies (e.g. Pinecone, Weaviate, pgvector, Vertex AI Search) in production or near production environments.
- Working knowledge of graph database technologies (e.g. Neo4j, Spanner Graph) and an understanding of when they are the right tool for the job.
- Hands on experience building and operating data lakes on Google Cloud Platform, with strong BigQuery skills including data modelling, partitioning, clustering, and cost management.
- Familiarity with the broader Google Cloud data ecosystem, including Cloud Dataflow, Cloud Composer, Pub/Sub and Looker.
- Proven ability to build reliable ETL/ELT pipelines using tooling such as Apache Airflow, dbt, and Google Dataflow.
- Good understanding of data integrity and consistency challenges in distributed systems, and practical experience implementing solutions to address them.
- Experience applying data quality checks and governance practices within data pipelines and storage layers.
- Comfortable contributing to technology discussions with product and engineering teams, and able to communicate technical decisions clearly.
- Takes real ownership of the quality and reliability of the data systems they build.
- Technically curious, keeping pace with developments in the data technology landscape.
- Collaborative and communicative, able to work effectively alongside product engineers, data scientists, and other technical stakeholders.
- Thorough and detail oriented in implementation, with the pragmatism to get things done in a fast moving environment.
We champion and welcome diversity in our workforce and ensure all job applicants receive equal and fair treatment, regardless of age, race, gender or gender identity, religion, sexual orientation, disability, or nationality.
We are not only committed to increasing the visibility and recognition of talent from under represented groups within our organisation, but the wider industry too.
At the end of the day, we make sure we take time to look after ourselves, each other, and the planet, because we're always stronger together.
ITG have a number of community groups (ERGs) available to employees which offer a safe space for like minded colleagues, shared interests to connect, socialise and check in with each other. These include Black ITGers Together, LGBTQ+ Together, Mens Health Together, Muslims Together, Neurodiversity Together, Working Parents and Carers Together and Women In Tech Together.
About This Role
Career insights for Database Architects positions