Data Engineer at Studyportals responsible for data pipelines and infrastructure. Join a team ensuring accurate and trustworthy data for analytics and business decisions.
Responsibilities
Design, build, and own production-grade ETL pipelines with strong data quality guarantees
Operate and evolve our AWS-based data platform (EMR, Redshift, Glue, zeroETL)
Run and improve core data tooling such as Airflow, dbt, Redshift, and Grafana
Ensure reliability in production by monitoring pipelines, debugging incidents, and preventing recurrence
Shape data architecture and infrastructure, contributing to Infrastructure-as-Code
Optimize performance and cost efficiency across the data stack as data volumes grow
Enable the business with trusted data that is well-documented, secure, and easy to use
Act as a go-to expert for data engineering topics and decisions across the company
Requirements
3+ years of hands-on data engineering experience in production cloud environments
Strong Python (PySpark) and SQL skills with a focus on clean, maintainable code
Proven experience building and operating reliable, scalable ETL pipelines
Solid experience with AWS and modern data stack tools such as Airflow, dbt, and Fivetran
Strong understanding of data modelling, incremental loading, and data quality practices
Experience hosting and maintaining data tools and infrastructure is a nice to have
Experience with streaming data processing and/or online tracking solutions is a nice to have
Experience optimizing warehouse performance and managing cloud costs is a nice to have
You take end-to-end ownership of systems running in production
You think in root causes and long-term solutions, not quick fixes
You communicate clearly in English in a highly international team
Benefits
Competitive salary including a bonus and pension.
32 vacation days (including bank holidays) you have the flexibility to take your holidays whenever you want (e.g., be off on Chinese New Year instead of King’s Day).
Reimbursement of your work-related travel costs.
Flexibility to work remotely 3 days per week, including a remote working allowance.
Possibility to work from abroad 2 times a year.
An annual personal development budget because personal growth is KEY!
A nice and fun office in the heart of Eindhoven; five minutes’ walk away from the station.
Senior Data Engineer responsible for building and optimizing data pipelines for banking analytics initiatives. Collaborating with data teams to ensure data quality and readiness for enterprise use.
Senior Data Engineer developing scalable data solutions on Databricks for analytics and operational workloads. Collaborating with cross - functional teams to modernize the data ecosystem.
Data Engineer focused on analytics and data pipeline development for network optimisation. Collaborating with teams to deliver high - quality data solutions with Python and SQL.
Senior Product Manager defining platform capabilities for Data Cloud in Salesforce. Collaborating with R&D teams while shaping product strategy for Data 360 integration.
Senior Data Engineer at Goodwin enhancing data platforms and fostering data - driven culture across teams. Collaborating with IT and Finance on technology solutions and data governance practices.
Director, Data Platform Design and Strategy at MedImpact leading data platform and AI innovations to enhance healthcare services. Overseeing enterprise projects and managing teams to meet strategic goals.
Data Engineer delivering AI - and data - driven solutions for Honeywell’s industrial customers. Architecting and implementing scalable data pipelines and platforms focused on IoT and real - time data processing.
Data Engineering Associate focusing on data quality control and management for distribution platform. Collaborates on large scale data projects to ensure data accuracy and availability for users.
Data Architect managing enterprise data platform built on Microsoft Fabric at Johnstone Supply. Leading architectural standards and collaborating with business and IT leaders for strategic data - driven insights.
AI/ML Engineer designing and refining prompts and workflows using large language models. Responsible for developing data pipelines and delivering scalable AI solutions in a hybrid work environment.