Data Engineer developing scalable data pipelines at Daikin Applied. Collaborating with teams to support advanced analytics and AI use cases using Databricks.
Responsibilities
Design, build, and maintain ETL/ELT pipelines to ingest, transform, curate and store data from multiple sources
Optimize data processing workflows for performance, reliability, and scalability
Implement real-time and batch data processing using technologies like Apache Spark, Kafka, and Databricks
Work with structured and unstructured data
Implement data validation, cleansing, and monitoring to ensure high-quality datasets
Implement data governance, security, and compliance policies (e.g., GDPR, CCPA)
Maintain metadata management, data lineage, and documentation for data assets
Deploy and manage data solutions on cloud platforms (Azure, Databricks)
Develop and maintain documentation, data models, and technical standards
Optimize query performance, cost efficiency, and storage utilization
Monitor, troubleshoot, and resolve issues in production data pipelines and environments
Stay current with the latest advancements in data engineering, cloud computing, and analytics technologies on the Databricks ecosystem
Partner with data analysts, and software engineers to support analytics initiatives
Requirements
Bachelor’s degree in Computer Science, Engineering, or a related field
8+ years of Data Engineering, with a strong understanding of cloud-based data solutions
At least 3 years hands-on experience building and delivering data products on Databricks
Proven experience in data engineering and pipeline development on Databricks
Hands-on expertise across the data lifecycle: ingestion, transformation, modelling, governance, and consumption
Deep expertise with the Databricks platform (SQL, Python and PySpark, Delta Lake, Unity Catalog, MLflow)
Strong SQL and Python skills for data processing and data manipulation
Strong problem-solving skills and an analytical mindset
Excellent verbal and written communication skills, with the ability to explain technical concepts to non-technical audiences
Extensive experience with data ingestion methodology including ADF
Proficiency in Python, SQL, or Scala for data processing
Experience with cloud data services (Azure Data Factory, Databricks)
Hands-on experience with big data frameworks (Databricks, Apache Spark)
Strong knowledge of data modeling, database optimization, and API-based data integration
Proficiency in designing and implementing the Medallion Architecture on Databricks
Experience with code repositories, CI/CD processes and release management
Work visa sponsorship is not available for this position
Benefits
Multiple medical insurance plan options + dental and vision insurance
401K retirement plan with employer contributions matching 100% of the first 3% of employee contributions and 50% on the next 2% of employee contributions
Company provided life insurance + optional employee paid voluntary life insurance, dependent life coverage and voluntary accident coverage
Short term and long-term disability
3 weeks of paid time off for new employees + 11 company paid holidays
Vacation accrues on a monthly basis, unless applicable federal, state and local law requires a faster accrual
Paid sick time in accordance of the federal, state and local law
Paid parental leave and tuition reimbursement after 6 months of continuous service
Senior Data Engineer developing scalable data pipelines and collaborating with cross - functional teams at Technis. Technical guidance in a hybrid work environment based in Lausanne, Switzerland.
Data Engineer designing and maintaining data pipelines for CIEE, a philanthropic institution supporting youth development. Collaborating with Data Analysts for data quality and reliability.
Senior Data Engineer responsible for designing and implementing data solutions at Harambee. Collaborating with various stakeholders to enhance technology supporting work - seekers' journeys.
Senior Consultant Data Engineer in a consultancy firm focusing on data engineering and platform development. Collaborating with diverse teams to deliver high - quality data solutions.
Senior Manager Data Engineer at Squarcle delivering technical leadership in data engineering and compliance with business objectives. Leading teams to optimize and develop data platforms for clients.
Data Engineer at Mobileye building robust data pipelines for data infrastructure. Collaborating with teams to deliver high - quality data solutions for dynamic environments.
Cloud Data Engineer at Shift, focusing on building and operating data pipelines on Azure for Australian SMEs. Collaborating across teams to enhance data integration and quality.
AI Systems & Data Engineer at HyperFi designing Databricks pipelines and managing AI systems. Collaborating on data engineering tasks and optimizing workflows in a flexible tech environment.
Engenheiro de Dados especializado em PowerBI e Lakehouse no Campos Thomaz Advogados. Foco em preparar dados para dashboards e estruturação de ambientes com Microsoft.
Data Platform Expert developing and maintaining data solutions for analysis and reporting at Magna Electronics. Collaborating with various teams to enhance data - driven decision making and insights.