Develop EL/ELT/ETL pipelines to make data available in BigQuery analytical data store from disparate batch, streaming data sources for the Business Intelligence and Analytics teams.
Work with on-prem data sources (Hadoop, SQL Server), understand the data model, business rules behind the data and build data pipelines (with GCP, Informatica) for one or more Ford Pro verticals. This data will be landed in GCP BigQuery.
Build cloud-native services and APIs to support and expose data-driven solutions.
Partner closely with our data scientists to ensure the right data is made available in a timely manner to deliver compelling and insightful solutions.
Design, build and launch shared data services to be leveraged by the internal and external partner developer community.
Building out scalable data pipelines and choosing the right tools for the right job. Manage, optimize and Monitor data pipelines.
Provide extensive technical, strategic advice and guidance to key stakeholders around data transformation efforts. Understand how data is useful to the enterprise.
Requirements
Bachelors Degree
3+ years of experience with SQL and Python
2+ years of experience with GCP or AWS cloud services; Strong candidates with 5+ years in a traditional data warehouse environment (ETL pipelines with Informatica) will be considered
3+ years of experience building out data pipelines from scratch in a highly distributed and fault-tolerant manner.
Comfortable with a broad array of relational and non-relational databases.
Proven track record of building applications in a data-focused role (Cloud and Traditional Data Warehouse)
Experience with GCP cloud services including BigQuery, Cloud Composer, Dataflow, CloudSQL, GCS, Cloud Functions and Pub/Sub.
Inquisitive, proactive, and interested in learning new tools and techniques.
Familiarity with big data and machine learning tools and platforms. Comfortable with open source technologies including Apache Spark, Hadoop, Kafka.
1+ year experience with Hive, Spark, Scala, JavaScript.
Strong oral, written and interpersonal communication skills
Comfortable working in a dynamic environment where problems are not always well-defined.
M.S. in a science-based program and/or quantitative discipline with a technical emphasis.
Senior Associate Data Engineer contributing to Travelers' analytics landscape by building and operationalizing data solutions. Collaborating with teams to ensure reliable data delivery across the enterprise.
Salesforce Data Engineer serving as a subject matter expert in the State of Tennessee. Designing scalable data pipelines and collaborating on cross - agency initiatives.
Data Engineer Senior responsible for building data architecture and optimizing pipelines for Business Intelligence. Collaborating with analysts to develop insights using Power BI and Azure technologies.
Principal Data Engineer driving modernization from legacy systems to cloud - native platforms at Mastercard. Architecting and developing ETL platforms with AI integration and establishing data - driven strategies.
Principal Data Engineer modernizing cloud - native platforms for AI - powered solutions at Mastercard. Leading teams to enhance data processing efficiency and reliability across global operations.
Data Engineer creating data pipelines for Santander's card transactions. Collaborating with an agile team in strategic projects involving Databricks and PySpark.
Data Engineer designing, implementing, and maintaining data pipelines at Sabiá Gaming. Focused on high - quality data access and integration for enhanced decision - making.
Quantitative Data Engineer developing data solutions and automations for MassMutual's investment management. Working with data orchestration tools within a collaborative team environment.
Senior Data Engineer designing and scaling data infrastructure for analytics, machine learning, and business intelligence in a software supply chain security company.