Founding Central Data Engineer at hyperexponential responsible for implementing data platform solutions. Collaborating with Central Data team to innovate and enhance data capabilities in insurance.
Responsibilities
Designing, building and owning hx’s Databricks lakehouse, including workspaces, clusters, Unity Catalog, medallion architecture, Delta tables, RBAC and cost controls, becoming the central source of truth for the company.
Implementing robust ELT pipelines (Spark, dbt or similar) with tests, documentation, CI and monitoring to deliver daily freshness and agreed hourly SLAs with ≤4h recovery for critical data.
Modelling cross-domain Gold entities and certified metrics across customers, usage, pricing, finance, people and GTM with clear grain, history and reconciliation paths that power all exec and board reporting.
Building hx’s first embedding, vector and retrieval layer over curated tables and documents, enabling live RAG and agent workflows such as premium explanations, version tracing and automated KPI summaries.
Establishing data contracts, data quality checks, audit and observability patterns so teams can trust, debug and self-serve confidently across the platform.
Acting as a founding partner to product, FP&A, people and GTM by unblocking ambiguous data problems, setting data standards from scratch and driving the company’s AI readiness.
Requirements
Built and run a Databricks-style platform with medallion layers, Unity Catalog, RBAC, SLAs, cost tuning and runbooks that served as a company’s primary data environment.
Delivered multi-domain models linking product/platform, finance, people and GTM data into coherent entities and certified metrics, including revenue reconciliation and historical tracking.
Created production-grade data pipelines using SQL, Python, Spark and Git workflows with tests, alerts, logging and documentation that elevated engineering quality across a team.
Built at least one real-world RAG or vector search system using embeddings, thoughtful chunking, metadata, evaluation and safe retrieval patterns.
Used data as software: robust CI, clear acceptance criteria, debugging discipline, observability, dev-to-prod promotion and well-maintained runbooks.
Served as an early or founding data hire, translating messy questions from senior leaders into clear designs, spotting systemic issues and raising the quality bar beyond your core remit.
Benefits
£5,000 training and conference budget for individual and group development.
25 days of holiday plus 8 bank holidays (33 days total).
Company pension scheme via Penfold.
Mental health support and therapy via Spectrum.life.
Data Engineering Academy focused on Snowflake and Databricks for professionals interested in expanding their technical capabilities. Fully remote with future office work in Monterrey or Saltillo after completion.
Senior Data Engineer at Intent HQ designing and scaling data platforms. Building high - impact intelligence from millions of customer insights with a focus on performance and reliability.
SAP Data Engineer supporting MERKUR GROUP's evolution into a data - driven company. Responsible for data integration, modeling, and collaboration with various departments in Group Finance.
Data Engineer at Booz Allen Hamilton organizing data and developing advanced technology solutions. Leading data engineering activities for mission - driven projects and mentoring multidisciplinary teams.
Senior Data Engineer at Bristol Myers Squibb developing scalable data pipelines for foundational products. Collaborating with data scientists and IT professionals to ensure data quality and accessibility.
Data Engineer II role focusing on developing and maintaining data pipelines for analytics. Collaborating with Data Science and Analytics teams to ensure data quality and reliability.
Senior Data Architecture Specialist designing and maintaining data integration solutions for Morgan Stanley. Involved in building data architecture and optimizing data storage using various technologies.
Lead Data Engineer responsible for building and maintaining the central HR data lake. Collaborating with analysts and business stakeholders for data - driven decision making.
Join GKN Aerospace as a Life Cycle Assessment & Sustainability Data Engineer. Conduct analyses of environmental impacts and manage sustainability data for a greener future.