Big Data Engineer developing applications for Synchrony’s Enterprise Data Lake within an Agile scrum team. Collaborating to deliver high-quality data ingestion and maintain data governance standards.
Responsibilities
Develop big data applications for Synchrony in Hadoop ecosystem
Participate in the agile development process including backlog grooming , coding, code reviews , testing and deployment
Work with team members to achieve business results in a fast paced and quickly changing environment
Work independently to develop analytic applications leveraging technologies such as: Hadoop, NoSQL, In-memory Data Grids, Kafka, Spark, Ab Initio
Provide data analysis for Synchrony’s data ingestion, standardization and curation efforts ensuring all data is understood from a business context
Identify enablers and level of effort required to properly ingest and transform data for the data lake.
Profile data to assist with defining the data elements, propose business term mappings, and define data quality rules
Work with the Data Office to ensure that data dictionaries for all ingested and created data sets are properly documented in data dictionary repository
Ensure the lineage of all data assets are properly documented in the appropriate enterprise metadata repositories
Assist with the creation and implementation of data quality rules
Ensure the proper identification of sensitive data elements and critical data elements
Create source-to-target data mapping documents
Test current processes and identify deficiencies
Investigate program quality to make improvements to achieve better data accuracy
Understand functional and non-functional requirement and prepare test data accordingly
Plan, create and manage the test case and test script
Identify process bottlenecks and suggest actions for improvement
Execute test script and collect test results
Present test cases, test results, reports and metrics as required by the Office of Agile
Perform other duties as needed to ensure the success of the team and application and ensure the team’s compliance with the applicable Data Sourcing, Data Quality, and Data Governance standards
Requirements
Bachelor's degree OR in lieu of Bachelor's degree, High School Diploma/ GED and minimum 2 years of Information Technology experience
Minimum of 1 year of Hands-on experience writing shell scripts , complex SQL queries, Hive scripts, Hadoop commands and Git
Ability to write abstracted, reusable code components
Programming experience in at least one of the following languages: Scala, Java or Python
Analytical mindset
Willingness and aptitude to learn new technologies quickly
Superior oral and written communication skills ; Ability to collaborate across teams of internal and external technical staff, business analysts, software support and operations staff.
Cloud Data Engineer at Regions designing, building, and maintaining data structures and pipelines. Collaborating on data initiatives, ensuring optimal architecture, working closely with technical partners.
Data Engineer in Veepee's Data Factory working on data ingestion pipelines and improving data quality. Collaborative environment utilizing Kubernetes, Python, Java, and modern data architectures.
Data Architect designing and maintaining enterprise data architecture at Envalior. Driving enterprise - wide impact ensuring scalability and reliability of systems, reporting, and AI initiatives.
Data Engineer role at Valmont focused on data analytics and technology for sustainable agricultural practices. Collaborating with cross - functional teams to enhance data management and analytics tools.
Senior Data Engineer at Barclays building and maintaining data pipelines and warehouses. Collaborating with data scientists and ensuring data accuracy, accessibility, and security.
Lead Data Engineer guiding a team in designing scalable data solutions for iKnowHow S.A. Overseeing development of data pipelines while collaborating with cross - functional teams.
Data Engineer at LPL Financial developing Python - based ETL pipelines. Collaborating with cross - functional teams to ensure reliable data delivery and optimizing pipeline performance.
Senior Data Engineer at Keyrus focusing on data solutions and projects to drive performance. Collaborating with teams globally to enhance data transformation and governance processes.
Data Engineer developing scalable data pipelines for ETL/ELT processes using GCP services. Collaborating with team members to optimize data workflows and ensure data integrity.
Data Governance Engineer in Fintech developing a formal cyber data governance framework. Collaborating with cyber security, analytics, and platform engineering teams on metadata and lineage capabilities.