Founding Central Data Engineer at hyperexponential responsible for implementing data platform solutions. Collaborating with Central Data team to innovate and enhance data capabilities in insurance.
Responsibilities
Designing, building and owning hx’s Databricks lakehouse, including workspaces, clusters, Unity Catalog, medallion architecture, Delta tables, RBAC and cost controls, becoming the central source of truth for the company.
Implementing robust ELT pipelines (Spark, dbt or similar) with tests, documentation, CI and monitoring to deliver daily freshness and agreed hourly SLAs with ≤4h recovery for critical data.
Modelling cross-domain Gold entities and certified metrics across customers, usage, pricing, finance, people and GTM with clear grain, history and reconciliation paths that power all exec and board reporting.
Building hx’s first embedding, vector and retrieval layer over curated tables and documents, enabling live RAG and agent workflows such as premium explanations, version tracing and automated KPI summaries.
Establishing data contracts, data quality checks, audit and observability patterns so teams can trust, debug and self-serve confidently across the platform.
Acting as a founding partner to product, FP&A, people and GTM by unblocking ambiguous data problems, setting data standards from scratch and driving the company’s AI readiness.
Requirements
Built and run a Databricks-style platform with medallion layers, Unity Catalog, RBAC, SLAs, cost tuning and runbooks that served as a company’s primary data environment.
Delivered multi-domain models linking product/platform, finance, people and GTM data into coherent entities and certified metrics, including revenue reconciliation and historical tracking.
Created production-grade data pipelines using SQL, Python, Spark and Git workflows with tests, alerts, logging and documentation that elevated engineering quality across a team.
Built at least one real-world RAG or vector search system using embeddings, thoughtful chunking, metadata, evaluation and safe retrieval patterns.
Used data as software: robust CI, clear acceptance criteria, debugging discipline, observability, dev-to-prod promotion and well-maintained runbooks.
Served as an early or founding data hire, translating messy questions from senior leaders into clear designs, spotting systemic issues and raising the quality bar beyond your core remit.
Benefits
£5,000 training and conference budget for individual and group development.
25 days of holiday plus 8 bank holidays (33 days total).
Company pension scheme via Penfold.
Mental health support and therapy via Spectrum.life.
Snowflake Data Engineer responsible for data pipelines and warehouses for enterprise analytics at Liberty Coca - Cola. Collaborating across business functions to ensure high data quality and performance.
Full - Stack Data Engineer designing and optimizing complex data solutions for automotive content. Collaborating with teams to enhance user experience across MOTOR's product lines.
Principal Data Engineer designing and evolving enterprise data platform. Collaborating with analytics teams to enable AI and data products at American Tower.
BI Data Engineer II supporting scalable Lakehouse data pipelines at Boston Beer Company. Collaborating with stakeholders to drive data ingestion and maintain enterprise data quality.
Senior Data Engineer at A Kube Inc responsible for building and maintaining data pipelines for product performance. Collaborating with product, engineering, and analytics teams to ensure data quality and efficiency.
Data Engineer engineering DUAL Personal Lines’ strategic data platforms for global insurance group. Providing technical expertise in data engineering and collaborating with internal teams for solution delivery.
Data Engineer role focused on creating and monitoring data pipelines in an innovative energy company. Collaborate with IT and departments to ensure quality data availability in a hybrid work environment.
SQL Migration Data Engineer at Auxo Solutions focusing on Azure SQL/Fabric Lakehouse migrations and building data pipelines. Collaborating on technical designs and data governance for modernization initiatives.