Data Architect needed to define and evolve data architecture supporting scientific compute at EIT. Collaborate and lead in large-scale research environments for transformative scientific challenges.
Responsibilities
**Your Responsibilities:**
__Data & Storage Architecture Collaboration__
Collaborate with research group technical leadership to define and continuously evolve the institute-wide data and storage architecture supporting large-scale scientific compute.
Define target-state data architectures for scientific data that enable EIT’s institutes to do their best work, balancing standardisation, flexibility, scalability, performance, resilience, and security across heterogeneous scientific data workloads.
Work hand-in-glove with other data teams within the Institute including Data Engineering (within AI and Robotics) and Enterprise Applications to ensure strategic and operational alignment across disciplines.
Translate organisational strategy and scientific priorities into coherent data architecture roadmaps for scientific compute.
____
__Standards, Schemas & Consistency__
Define and own institute-wide data standards that are utilised by our scientists and developers, including schemas, metadata models, naming conventions, and configuration baselines.
Ensure that consistent standards, schemas, and configuration settings are used across all scientific programmes, wherever appropriate.
Balance standardisation with scientific flexibility, providing clear, governed patterns for extension rather than divergence.
____
__Large-Scale Data Integration & AI Enablement__
Architect approaches for integrating data across scientific programmes into unified, high-quality datasets.
Enable the creation of very large datasets suitable for advanced analytics, machine learning, and large language model training.
Work closely with scientific compute, AI, and platform teams to ensure data architectures are optimised for large-scale downstream consumption.
____
__Data Governance, Classification & Compliance__
Be accountable for aligning data and storage operational standards with data classification models defined by the Institute’s Data Protection Officer (DPO) and other data specialists.
Translate governance, privacy, and security requirements into clear, practical architectural and operational standards.
Ensure data is handled appropriately throughout its lifecycle, including ingestion, storage, access, sharing, retention, and deletion.
____
__Collaboration & Influence__
Act as a trusted partner to scientific compute leaders across programmes, engaging deeply with their requirements, constraints, and research priorities.
Lead through influence rather than mandate, collectively defining shared schemas and standards that programmes commit to and adopt.
Support programmes with the operationalisation of agreed standards, ensuring they are embedded into delivery pipelines and day-to-day practices.
____
__Operationalisation & Enablement__
Ensure data architecture standards move beyond definition into implementation and sustained operation.
Provide guidance, reference architectures, and hands-on support to programme teams adopting shared data and storage standards.
Work alongside platform, DevOps, and operations teams to embed standards into tooling, automation, and operational processes.
Distinguished Engineer driving design and architecture of mission - critical Data Platforms at Capital One. Leading technical strategies to enhance data discovery and governance for enterprise - wide AI readiness.
Principal Software Engineer at Clari + Salesloft developing enterprise - grade AI - driven applications for revenue intelligence with a dynamic team in India.
Data Engineer responsible for ingestion pipelines and data quality on AI - driven marketing platform. Collaborating with data teams to ensure accuracy and performance of data systems.
Data Architect designing and governing data foundations for analytics and AI applications at Clio. Collaborating cross - functionally to develop high - quality data models and standards.
Software Engineer at Warner Music Group developing an innovative Data Platform for the music industry. Collaborating with dynamic teams to enhance music data processing and delivery.
Data Engineer role specializing in Azure & Snowflake at InfoCentric. Leading design and delivery of enterprise - scale data platforms for large organizations.
Principal Data Architect at PointClickCare ensuring coherent and scalable data architecture. Driving unified data direction while collaborating with Engineering Architecture team for AI enablement.
Data Engineer Tech Lead developing data solutions at Carelon. Leading a cross - functional team to optimize data workflows and maintain data integrity.
Lead Data Engineer responsible for evolving Manna’s data infrastructure for drone delivery. Overseeing data architecture and analytics while building scalable data pipelines.
Data Engineer designing, implementing, and optimizing data pipelines for DeepLight AI. Collaborating closely with a multidisciplinary team to analyze large - scale data.