Data Engineer creating data pipelines for Santander's card domain. Collaborating with an agile team on strategic projects and leveraging Databricks and PySpark expertise.
Responsibilities
The Data Engineer will work with the Card team’s Data Engineering group to create data pipelines for ingesting and exposing card-domain data in Santander Brazil’s Corporate Data Lake.
The person will work in an agile team on a strategic area project and should have experience with Databricks and PySpark.
Requirements
Databricks proficiency: Experience working with Apache Spark on Databricks, including building and optimizing data pipelines.
PySpark, Python and Kedro experience: Strong programming skills in PySpark and Python and experience with Kedro to develop, debug and maintain data transformation code.
Batch and streaming data processing: Knowledge of batch and streaming (messaging) data processing, with the ability to design, implement and maintain data processing pipelines.
DevOps knowledge: Familiarity with Jenkins for continuous integration and continuous delivery (CI/CD), as well as automation of deployment tasks and pipeline management.
Git: Proficiency with Git for source code version control and effective team collaboration.
Agile methods: Understanding of agile principles and practices such as Kanban and Scrum for effective collaboration and project management.
Orchestration (e.g., Control‑M or others): Knowledge of workflow orchestration tools, important for scheduling and controlling workflows.
Microsoft Azure knowledge: Experience with key Microsoft Azure data services, including Azure Databricks, Azure Data Factory and Azure Storage.
AWS knowledge: Experience with key AWS services such as Aurora PostgreSQL, CloudWatch, Lambda and S3.
On‑Premises environments (Cloudera) experience: Previous experience with the Cloudera platform or other on‑premises big data solutions, including Hadoop, HBase and Hive, is desirable.
Object‑oriented development knowledge: Familiarity with Java is helpful (not required to write code, but to interpret it).
Optional certifications: AZ‑900 (Microsoft Azure Fundamentals) and DP‑900 (Microsoft Azure Data Fundamentals) are preferred and demonstrate solid knowledge of the Azure platform and data fundamentals.
Benefits
Bradesco Health Plan (30% co-payment)
Bradesco Dental Plan (no employee contribution)
Life Insurance
Wellhub (Gympass)
Childcare allowance
Allowance for children with special needs
Payroll‑deductible loan
Private pension
Pet plan
SESC benefits
Conexa telemedicine
Cost allowance
Meal / Food voucher
Multi‑benefits card
Medical plan upgrade
DIFFERENTIALS:
We are a socially responsible employer: extended maternity and paternity leave
INMaterna Program: support program for pregnant employees
Newborn welcome kit and the book "It Happened When I Was Born"
Professional development: courses available through the internal university
100% remote or hybrid, depending on project applicability.
Data Engineer managing and organizing datasets for AI models at Walaris, developing AI - driven autonomous systems for defense and security applications.
Data Engineer designing and maintaining data pipelines at Black Semiconductor. Collaborating with process, equipment, and IT teams to support manufacturing analytics and decision - making.
Junior Data Engineer role focusing on Business Intelligence and Big Data at Avanade. Collaborating on data analysis and SQL queries in a supportive learning environment.
GCP Data Engineer designing and developing data processing modules for Ki, an algorithmic insurance carrier. Working closely with multiple teams to optimize data pipelines and reporting.
Data Engineer at Securian Financial optimizing scalable data pipelines for AI and advanced analytics. Collaborating with teams to deliver secure and accessible data solutions.
IT Data Engineering Co‑Op at BlueRock Therapeutics supports development of scientific data systems. Collaboration on data workflows and foundational AWS data engineering tasks.
Data Engineer I building and operationalizing complex data solutions for Travelers' analytics using Databricks. Collaborating within teams to educate end users and support data governance.
Data Engineer shaping modern data architecture to drive golf’s digital transformation. Collaborating with teams to enhance data pipelines and insights for customer engagement and revenue growth.
Staff Data Engineer overseeing complex data systems for CITY Furniture. Responsible for architecting and optimizing data ecosystems in a hybrid work environment.
Data Engineer strengthening data platform team at Samba TV to improve data analytics and reporting capabilities. Building on AWS, Databricks, BigQuery, and Snowflake technology.