About the role

  • Senior PySpark Developer managing data processing pipelines for a data project focusing on cloud integration and governance. Collaborating with teams to ensure data quality and efficiency.

Responsibilities

  • Develop and maintain data pipelines using PySpark
  • Perform transformations on large volumes of data in distributed environments
  • Integrate data flows with Object Storage solutions
  • Work with Oracle databases for ingestion, querying, and data persistence
  • Ensure quality, consistency, and integrity of processed data
  • Optimize Spark job performance and processing routines
  • Collaborate with data, architecture, and business teams to understand requirements
  • Document processes, data flows, and technical decisions

Requirements

  • Strong experience with Python applied to data engineering
  • Advanced experience with Apache Spark and PySpark
  • Experience in large-scale data transformation and processing projects
  • Experience with Object Storage (reading, writing, and organizing data)
  • Knowledge of Oracle Database (queries, basic modeling, integration)
  • Experience building end-to-end data pipelines

Benefits

  • Availability to work on weekends, occasionally for go-lives

Job title

Senior PySpark Developer

Job type

Experience level

Senior

Salary

Not specified

Degree requirement

Bachelor's Degree

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job