Hybrid Data Engineer

Posted 3 days ago

Apply now

About the role

  • Ingest data from a variety of sources such as Azure SQL DB, Google Analytics, Google Play Store, Apple App Store, Salesforce, and others.
  • Develop and optimize ETL/ELT pipelines to transform data from CSV, JSON, SQL tables, and APIs into usable formats.
  • Work with REST APIs to pull data from various external sources and integrate it into our data ecosystem.
  • Design and implement efficient data transformation processes to cleanse, aggregate, and enrich data.
  • Apply industry best practices for data modeling to ensure scalability, performance, and data integrity.
  • Collaborate with data analysts and data scientists to provide clean, high-quality datasets for reporting and analysis.
  • Utilize Databricks for data processing, transformation, and orchestration tasks.
  • Manage and optimize Databricks clusters for performance, reliability, and cost-effectiveness.
  • Implement Databricks workflows to automate and streamline data pipelines.
  • Use Unity Catalog for data governance and metadata management, ensuring compliance and data access control.

Requirements

  • 5+ years of hands-on experience in data engineering or a related field.
  • Proven experience with Databricks and Databricks workflows, including cluster management and data pipeline orchestration.
  • Strong experience in data ingestion from SQL databases (Azure SQL DB), APIs (Google Analytics, Google Play Store, Apple App Store, Salesforce), and file-based sources (CSV, JSON).
  • Proficiency in SQL for data manipulation and transformation.
  • Experience with Python or Scala for writing and managing data workflows.
  • Working knowledge of REST APIs for data integration.
  • Experience in data transformation using Apache Spark, Delta Lake, or similar technologies.
  • Knowledge of cloud platforms such as Azure, with a focus on Azure SQL DB.
  • Familiarity with Unity Catalog for metadata management and governance.
  • Understanding of data architecture, data pipelines, and the ETL/ELT process.
  • Experience in data modeling, optimizing queries, and working with large datasets.
  • Familiar with data governance, metadata management, and data access controls.
  • Knowledge of Apache Kafka or other real-time streaming technologies (optional).
  • Experience with Data Lake or Data Warehouse technologies (optional).
  • Familiarity with additional data transformation tools such as Apache Airflow or dbt (optional).
  • Understanding of machine learning workflows and data pipelines (optional).

Job title

Data Engineer

Job type

Experience level

Mid levelSenior

Salary

Not specified

Degree requirement

Bachelor's Degree

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job