Data Architect needed to define and evolve data architecture supporting scientific compute at EIT. Collaborate and lead in large-scale research environments for transformative scientific challenges.
Responsibilities
**Your Responsibilities:**
__Data & Storage Architecture Collaboration__
Collaborate with research group technical leadership to define and continuously evolve the institute-wide data and storage architecture supporting large-scale scientific compute.
Define target-state data architectures for scientific data that enable EIT’s institutes to do their best work, balancing standardisation, flexibility, scalability, performance, resilience, and security across heterogeneous scientific data workloads.
Work hand-in-glove with other data teams within the Institute including Data Engineering (within AI and Robotics) and Enterprise Applications to ensure strategic and operational alignment across disciplines.
Translate organisational strategy and scientific priorities into coherent data architecture roadmaps for scientific compute.
____
__Standards, Schemas & Consistency__
Define and own institute-wide data standards that are utilised by our scientists and developers, including schemas, metadata models, naming conventions, and configuration baselines.
Ensure that consistent standards, schemas, and configuration settings are used across all scientific programmes, wherever appropriate.
Balance standardisation with scientific flexibility, providing clear, governed patterns for extension rather than divergence.
____
__Large-Scale Data Integration & AI Enablement__
Architect approaches for integrating data across scientific programmes into unified, high-quality datasets.
Enable the creation of very large datasets suitable for advanced analytics, machine learning, and large language model training.
Work closely with scientific compute, AI, and platform teams to ensure data architectures are optimised for large-scale downstream consumption.
____
__Data Governance, Classification & Compliance__
Be accountable for aligning data and storage operational standards with data classification models defined by the Institute’s Data Protection Officer (DPO) and other data specialists.
Translate governance, privacy, and security requirements into clear, practical architectural and operational standards.
Ensure data is handled appropriately throughout its lifecycle, including ingestion, storage, access, sharing, retention, and deletion.
____
__Collaboration & Influence__
Act as a trusted partner to scientific compute leaders across programmes, engaging deeply with their requirements, constraints, and research priorities.
Lead through influence rather than mandate, collectively defining shared schemas and standards that programmes commit to and adopt.
Support programmes with the operationalisation of agreed standards, ensuring they are embedded into delivery pipelines and day-to-day practices.
____
__Operationalisation & Enablement__
Ensure data architecture standards move beyond definition into implementation and sustained operation.
Provide guidance, reference architectures, and hands-on support to programme teams adopting shared data and storage standards.
Work alongside platform, DevOps, and operations teams to embed standards into tooling, automation, and operational processes.
Founding Staff Data Engineer building and leading data engineering team for AI - driven art valuation platform. Establishing architecture and standards for data systems and pipelines.
Senior Data Engineer responsible for developing, maintaining ETL processes and integrating data solutions. Collaborating with teams on data quality and cloud migration initiatives.
Data Engineer optimizing data architectures and pipelines at Nexu. Focused on building reliable and efficient data flows while collaborating with cross - functional teams.
Senior Software Engineer designing and maintaining scalable data solutions for restaurant tech industry at SpotOn. Collaborating with cross - functional teams to enhance reporting and analytics platforms.
Engineering Data Coordinator leading a data engineering team in Azure and Databricks at Deroyque. Focusing on project management, quality assurance, and team development based in Campinas.
Data Migration Specialist managing ongoing Salesforce data quality initiatives for Abby Care. Executing and validating data migrations while ensuring data accuracy.
Lead Data Engineer overseeing and managing the Data Engineering team. Developing ETL pipelines and ensuring data integrity within Cloud (Azure) infrastructure.
Data Engineer designing and optimizing data solutions for Qualco Intelligent Finance. Focus on data integrity, consistency, and reusability in analytics deliverables within a hybrid environment.
Data Architect leading data architecture and design for LifeByte's technology ecosystem. Collaborating with multiple teams to ensure robust data governance, compliance, and innovative strategies.