About the role

  • Design, build, and maintain high-performance ETL/ELT pipelines using PySpark on AWS.
  • Develop automated ingestion, transformation, and validation workflows for large structured and semi-structured datasets.
  • Optimize Spark jobs for performance, scalability, and cost efficiency.
  • Build and manage data pipelines that load into Snowflake using PySpark, Snowpipe, and external stages.
  • Create and maintain Snowflake objects including: Databases, schemas, tables, Virtual warehouses, Internal/external stages, file formats, Streams, Tasks, Dynamic Tables.
  • Implement Snowpipe for continuous or incremental ingestion.
  • Apply Snowflake optimization techniques (clustering, micro-partitioning, query profiling, etc.).
  • Work with AWS services such as S3, IAM, Lambda, CloudWatch, and EventBridge for data ingestion and automation.
  • Implement event-driven ingestion using SNS/SQS or other AWS-native triggers.

Requirements

  • 5+ years of experience with PySpark, including performance tuning, DataFrames, Spark SQL, and distributed data processing.
  • 3+ years of hands-on experience with Snowflake, including Snowpipe, stages, tasks, streams, and performance optimization.
  • Strong experience building data pipelines on AWS.
  • Strong SQL skills with the ability to write optimized, complex queries.
  • Solid understanding of ETL/ELT concepts, data warehousing, and modern data architecture.

Job title

AWS Data Engineer

Job type

Experience level

Mid levelSenior

Salary

Not specified

Degree requirement

Bachelor's Degree

Location requirements

Report this job

See something inaccurate? Let us know and we'll update the listing.

Report job