Ingest data from a variety of sources such as Azure SQL DB, Google Analytics, Google Play Store, Apple App Store, Salesforce, and others.
Develop and optimize ETL/ELT pipelines to transform data from CSV, JSON, SQL tables, and APIs into usable formats.
Work with REST APIs to pull data from various external sources and integrate it into our data ecosystem.
Design and implement efficient data transformation processes to cleanse, aggregate, and enrich data.
Apply industry best practices for data modeling to ensure scalability, performance, and data integrity.
Collaborate with data analysts and data scientists to provide clean, high-quality datasets for reporting and analysis.
Utilize Databricks for data processing, transformation, and orchestration tasks.
Manage and optimize Databricks clusters for performance, reliability, and cost-effectiveness.
Implement Databricks workflows to automate and streamline data pipelines.
Use Unity Catalog for data governance and metadata management, ensuring compliance and data access control.
Requirements
5+ years of hands-on experience in data engineering or a related field.
Proven experience with Databricks and Databricks workflows, including cluster management and data pipeline orchestration.
Strong experience in data ingestion from SQL databases (Azure SQL DB), APIs (Google Analytics, Google Play Store, Apple App Store, Salesforce), and file-based sources (CSV, JSON).
Proficiency in SQL for data manipulation and transformation.
Experience with Python or Scala for writing and managing data workflows.
Working knowledge of REST APIs for data integration.
Experience in data transformation using Apache Spark, Delta Lake, or similar technologies.
Knowledge of cloud platforms such as Azure, with a focus on Azure SQL DB.
Familiarity with Unity Catalog for metadata management and governance.
Understanding of data architecture, data pipelines, and the ETL/ELT process.
Experience in data modeling, optimizing queries, and working with large datasets.
Familiar with data governance, metadata management, and data access controls.
Knowledge of Apache Kafka or other real-time streaming technologies (optional).
Experience with Data Lake or Data Warehouse technologies (optional).
Familiarity with additional data transformation tools such as Apache Airflow or dbt (optional).
Understanding of machine learning workflows and data pipelines (optional).
Senior Associate Data Engineer contributing to Travelers' analytics landscape by building and operationalizing data solutions. Collaborating with teams to ensure reliable data delivery across the enterprise.
Salesforce Data Engineer serving as a subject matter expert in the State of Tennessee. Designing scalable data pipelines and collaborating on cross - agency initiatives.
Data Engineer Senior responsible for building data architecture and optimizing pipelines for Business Intelligence. Collaborating with analysts to develop insights using Power BI and Azure technologies.
Principal Data Engineer driving modernization from legacy systems to cloud - native platforms at Mastercard. Architecting and developing ETL platforms with AI integration and establishing data - driven strategies.
Principal Data Engineer modernizing cloud - native platforms for AI - powered solutions at Mastercard. Leading teams to enhance data processing efficiency and reliability across global operations.
Data Engineer creating data pipelines for Santander's card transactions. Collaborating with an agile team in strategic projects involving Databricks and PySpark.
Data Engineer designing, implementing, and maintaining data pipelines at Sabiá Gaming. Focused on high - quality data access and integration for enhanced decision - making.
Quantitative Data Engineer developing data solutions and automations for MassMutual's investment management. Working with data orchestration tools within a collaborative team environment.
Senior Data Engineer designing and scaling data infrastructure for analytics, machine learning, and business intelligence in a software supply chain security company.