Senior Data Engineer designing scalable ETL data pipelines using Databricks for a software consulting company. Collaborating with teams to implement robust data solutions in diverse business environments.
Responsibilities
Design and implementation of robust, scalable, and high-performance ETL/ELT data pipelines using PySpark/Scala and Databricks SQL on the Databricks platform;
Strong expertise in implementing and optimizing the Medallion Architecture (Bronze, Silver, Gold) using Delta Lake, ensuring data quality, consistency, and historical tracking.
Efficient implementation of the Lakehouse architecture on Databricks, combining best practices from traditional Data Warehousing and Data Lake paradigms;
Optimization of Databricks clusters, Spark operations, and Delta tables (e.g. Z-Ordering, compaction, query tuning) to reduce latency and compute costs;
Design and implementation of real-time and near-real-time data processing solutions using Spark Structured Streaming and Delta Live Tables (DLT);
Implementation and administration of Unity Catalog for centralized data governance, fine-grained security (row- and column-level security), and end-to-end data lineage;
Definition and implementation of data quality standards and validation rules (e.g. using DLT or Great Expectations) to ensure data integrity and reliability;
Development and management of complex workflows using Databricks Workflows (Jobs) or external orchestration tools such as Azure Data Factory or Airflow to automate data pipelines;
Integration of Databricks pipelines into CI/CD processes using Git, Databricks Repos, and Databricks Bundles;
Close collaboration with Data Scientists, Analysts, and Architects to translate business requirements into optimal technical solutions;
Providing technical mentorship to junior engineers and promoting engineering best practices across the team.
Requirements
Proven, expert-level experience across the full Databricks ecosystem, including Workspace management, cluster configuration, notebooks, and Databricks SQL.
In-depth knowledge of Spark architecture (RDDs, DataFrames, Spark SQL) and advanced performance optimization techniques;
Strong expertise in implementing and managing Delta Lake features, including ACID transactions, Time Travel, MERGE operations, OPTIMIZE, and VACUUM;
Advanced/expert proficiency in Python (PySpark) and/or Scala (Spark);
Expert-level SQL skills and strong experience with data modeling approaches (Dimensional Modeling, 3NF, Data Vault);
Solid hands-on experience with a major cloud platform (AWS, Azure, or GCP), with a strong focus on cloud storage services (S3, ADLS Gen2, GCS) and networking fundamentals.
**
**Nice to have**
Practical experience implementing and administering Unity Catalog for centralized governance and fine-grained access control;
Hands-on experience with Delta Live Tables (DLT) and Databricks Workflows for building and orchestrating data pipelines;
Basic understanding of MLOps concepts and hands-on experience with MLflow to support collaboration with Data Science teams;
Experience with Terraform or equivalent Infrastructure as Code (IaC) tools;
Databricks certifications (e.g. Databricks Certified Data Engineer Professional) are considered a significant advantage;
Bachelor’s degree in Computer Science, Engineering, Mathematics, or a related technical field;
5+ years of experience in Data Engineering, including at least 3+ years working with Databricks and Apache Spark at scale.
Benefits
Premium medical package
Lunch Tickets & Pluxee Card
Bookster subscription
13th salary and yearly bonuses
Enterprise job security with a startup mentality (diverse & engaging environment, international exposure, flat hierarchy) under the stability of a secure multinational
A supportive culture (we value ownership, autonomy, and healthy work-life balance) with great colleagues, team events and activities
Flexible working program and openness to remote work
Collaborative mindset – employees shape their own benefits, tools, team events and internal practices
Diverse opportunities in Software Development with international exposure
Flexibility to choose projects aligned with your career path and technical goals
Access to leading learning platforms, courses, and certifications (Pluralsight, Udemy, Microsoft, Google Cloud)
Career growth & learning – mentorship programs, certifications, professional development opportunities, and above-market salary
Health Data Engineer organizing big data and implementing data engineering activities for mission - driven healthcare projects at Booz Allen. Collaborating with a multidisciplinary team in a fast - paced environment.
Data Engineer responsible for developing and maintaining data engineering solutions at Ørsted. Collaborating with stakeholders and modernising database architecture for future needs.
Senior Azure Data Engineer at Accellor enhancing data capabilities for enterprise clients using Azure technologies. Building solutions to transform business processes through robust data engineering practices.
Data Architect at Evertec designing and implementing data architectures in Sao Paulo. Collaborating with cross - functional teams and ensuring data governance and quality practices.
Data Engineer Senior creating scalable data pipelines and deploying ML models. Collaborating within a passionate team at Norsys focusing on IT engineering and consulting.
Data Engineer at Cerium Networks, responsible for designing and implementing data solutions for clients. Focus on advanced architectures and Generative AI integration, collaborating with sales and technical teams.
Data Engineer responsible for architecting and building data solutions for clients at Cerium Networks. Collaborating with teams to leverage advanced data architectures and ensure client satisfaction.
Data Engineer at Cerium Networks designing and optimizing data solutions for clients. Employing advanced architectures and providing consulting services while supporting client integrations.
Data Engineer designing and optimizing data solutions for Cerium clients. Leveraging advanced architectures for data estates and providing strategic insights through analytics.