Data Engineer designing and optimizing data pipelines using Databricks and Google Cloud Platform. Collaborating with analysts and scientists to deliver high-quality data products.
Responsibilities
Design, build, and maintain ETL/ELT pipelines using Databricks (PySpark, Delta Lake).
Optimize pipelines for performance, cost efficiency, and scalability within GCP.
Develop batch and streaming data processes using Spark Streaming and related technologies.
Implement data solutions leveraging GCP services such as BigQuery, Cloud Storage, Dataflow, Cloud Composer, and Vertex AI integrations.
Apply best practices for cloud security, IAM configuration, monitoring, and cost management.
Build and maintain data models, including dimensional modeling and data vault structures.
Implement data quality frameworks, validation rules, and automated testing.
Managing data versioning, governance, and lineage using tools such as Unity Catalog or GCP Data Catalog.
Partner with cross-functional teams to gather requirements and translate them into technical designs.
Provide technical guidance and influence engineering best practices across the team.
Contribute to documentation, architectural diagrams, and knowledge sharing.
Requirements
Data Engineer or similar role.
Strong hands-on experience with Databricks, including: PySpark/Spark, Delta Lake, Databricks workflows/jobs
Proficiency with GCP: BigQuery, Cloud Storage, Dataflow or Dataproc
Strong coding skills in Python and SQL
Solid understanding of distributed systems, data warehousing, and data architecture principles.
Experience with CI/CD tools (GitHub, GitLab, Azure DevOps, or similar).
Benefits
Paid time off based on employee grade (A-F), defined by policy: Vacation: 12-25 days, depending on grade
Company paid holidays
Personal Days
Sick Leave
Medical, dental, and vision coverage (or provincial healthcare coordination in Canada)
Retirement savings plans (e.g., 401(k) in the U.S., RRSP in Canada)
Life and disability insurance
Employee assistance programs
Other benefits as provided by local policy and eligibility
Job title
Data Engineer, Databricks, Google Cloud Platform (GCP)
Data Engineer responsible for building scalable data infrastructure that supports data - driven decisions. Collaborating with team to maintain systems and unlock data value for organizations.
Associate Data Engineer supporting privacy engineering controls and executing privacy impact assessments in a financial services company. Collaborating across business units to ensure alignment with privacy regulations.
Data Engineer at CVS Health optimizing data pipelines and analytical models. Driving data - driven decisions with healthcare data for improved business outcomes.
Senior Data Engineer at CVS Health developing robust data pipelines for healthcare data. Collaborating with teams to provide actionable insights and integrate them with consumer touchpoints.
Senior Data Engineer supporting AI - enabled financial compliance initiative with data pipelines and ingestion processes. Collaborating with diverse teams in a mission - critical regulated environment.
Data Architect leading the definition and construction of cloud data architecture for Kyndryl. Participating in significant technological modernization initiatives, focusing on Google Cloud Platform.
Senior Data Engineer driving data intelligence requirements and scalable data solutions for a global consulting firm. Collaborating across functions to enhance Microsoft architecture and analytics capabilities.
Experienced AI Engineer designing and building production - grade agentic AI systems using generative AI and large language models. Collaborating with data engineers, data scientists in a tech - driven company.
Intermediate Data Engineer designing and building data pipelines for travel industry data management. Collaborating across teams to ensure reliable data for analytics and reporting.
Data Engineer managing and organizing datasets for AI models at Walaris, developing AI - driven autonomous systems for defense and security applications.