About the role

  • Principal Data Pipeline Lead at SS&C overseeing development of scalable data pipelines. Leading a small team and providing technical guidance for modern data platform integration.

Responsibilities

  • Lead the development of batch and real-time data pipelines on top of a modern data platform
  • Design and build scalable ingestion and transformation pipelines
  • Mentor a small team of engineers
  • Collaborate with platform engineering team to build pipelines
  • Implement CDC pipelines using Debezium and Kafka
  • Build streaming pipelines using Kafka and Apache Flink
  • Develop transformation workflows using Python, Spark / PySpark, and Airflow
  • Ingest data from DB2 replication streams
  • Process legacy fixed-width and CSV data feeds
  • Integrate API-based data sources
  • Store and manage data using Apache Iceberg and Parquet
  • Enable analytics through Trino and StarRocks

Requirements

  • 8+ years building data platforms or large-scale data pipelines
  • Strong programming experience in Python
  • Experience with Spark / PySpark
  • Experience building pipelines with Apache Airflow
  • Experience with Kafka-based streaming architectures
  • Experience implementing CDC pipelines (Debezium or similar)
  • Experience with Apache Flink or other streaming frameworks
  • Experience with Parquet and modern table formats such as Apache Iceberg
  • Experience with distributed query engines such as Trino, Presto, or StarRocks
  • Experience integrating data from heterogeneous or legacy systems
  • Experience leading or mentoring engineers

Benefits

  • Competitive salary
  • Opportunities for increased leadership scope as the team expands

Job title

Principal Data Pipeline Lead

Job type

Experience level

Senior

Salary

Not specified

Degree requirement

No Education Requirement

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job