Azure Data Engineer developing scalable data pipelines and collaborating with data teams for EU institutions. Working remotely from EU locations with a focus on high-quality data management.
Responsibilities
Develop, deploy, and maintain scalable and incremental data pipelines from REST APIs and databases using Python, PySpark, Azure Synapse, Knime, SQL, and ETL tools to ingest, transform, and prepare data.
Process and transform complex JSON and GIS data into structured datasets optimized for analysis and reporting.
Load, organize, and manage data in Azure Data Lake Storage and Microsoft Fabric OneLake, ensuring accessibility, performance, and efficient storage using lakehouse and Delta Lake patterns.
Document ETL processes, metadata definitions, data lineage, and technical specifications to ensure transparency and reusability.
Collaborate with data analysts, BI developers, and business stakeholders to understand data requirements and deliver reliable, well-documented datasets aligned with organizational needs.
Implement data quality checks, logging, monitoring, and automated incremental load mechanisms within data pipelines to support maintainability, observability, and troubleshooting.
Requirements
Bachelor degree - EQF 6
Ability to understand, speak and write English (C1/C2), French (B2) will be an advantage
Excellent knowledge of data engineering tools Azure Synapse Analytics, Microsoft Fabric, PySpark and Python.
Excellent knowledge of working with REST APIs, including ingestion and parsing of JSON and GIS data
Excellent knowledge of Azure data lake storage and Oracle database
Experience designing incremental loads, CDC processes, and automated schema evolution
Ability to implement robust data quality checks, logging, and monitoring in ETL processes
Ability to document ETL workflows, metadata, and technical specifications clearly and consistently
Familiarity with DevOps and version control best practices.
Experience with CI/CD pipelines
Experience working in an Agile and Scrum framework
Analysis and problem solving skills
Ability to participate in technical meetings and good communication skills
Ability to participate in multilingual meetings
Ability to work in multi-cultural environment, on multiple large projects; Excellent Team Player
Specific expertise (mandatory) at least 5 years of excellent knowledge in Azure Data Lake Storage, Microsoft Fabric OneLake, and Oracle databases
at least 5 years of excellent expertise in developing data pipelines from REST APIs and on integration (such as Azure Synapse, PySpark, Microsoft Fabric, Python, SQL, KNIME)
at least 5 years of excellent expertise in processing JSON and GIS data
Certificates (mandatory) Microsoft Azure Data Engineer Associate
Following certifications are a plus: Microsoft Certified: Azure Solutions Architect Expert
Microsoft Certified: Azure Developer Associate
Microsoft Certified: Azure Database Administrator Associate
Data Engineer/Analyst maintaining and improving data infrastructure for Braiins. Collaborating with technical and business teams to ensure reliable data flows and insights.
Medior Data Engineer handling Azure migrations for a major urban mobility client. Focused on data pipeline development and ensuring platform reliability with cutting - edge technologies.
Developing ML and computer vision solutions for cutting - edge autonomous vehicle dataset pipeline at Mobileye. Collaborating across teams for data curation and advanced perception algorithms.
Data Migration Lead in a hybrid role managing data migration for a major transformation programme in the media sector. Collaborating with various teams to ensure data integrity and successful migration.
Consultant ML & DataOps at Smile integrating data science projects for major clients. Designing MLOps solutions and enhancing data governance in a collaborative environment.
Data Engineer developing and maintaining data pipelines for Coolbet’s analytical services. Working within an Agile framework to ensure data reliability and efficiency.
API Data Engineer developing innovative data - driven solutions and advancing data architecture for AI Control Tower. Building and integrating APIs and data pipelines to support organizational needs.
Journeyman Data Architect supporting Leidos' enterprise data and analytics program for the Department of War. Collaborating on solutions for data architecture, cloud environments, and governance.
Senior Software Engineer developing backend services and data infrastructure for integrated products at Booz Allen. Collaborating with a small elite team to deliver reliable and scalable services.
AWS Streaming Data Engineer developing software and systems in a fast, agile environment. Utilizing experience with real - time data ingestion and processing systems across distributed environments.