Data Engineer at Kyndryl designing and maintaining data pipelines using AWS and Python. Optimizing ingestion, transformation workflows, and cloud solutions for large-scale data environments.
Responsibilities
designing, developing, and maintaining large‑scale data pipelines using Python, PySpark, and SQL
building and optimizing data ingestion, transformation, and processing workflows on AWS using services such as Lambda, Glue, EMR, S3, Athena, DynamoDB, Step Functions, MWAA, EventBridge, SNS/SQS, and Kinesis
implementing secure, scalable, and reliable cloud solutions aligned with AWS best practices through IAM, CloudWatch, CloudTrail, and Secrets Manager
working extensively with modern table formats like Apache Iceberg, Delta Lake, or Apache Hudi
developing and tuning advanced SQL queries, including Teradata SQL
contributing to scalable, reliable, and cost‑efficient Data Lake and Lakehouse architectures
designing event‑driven and serverless data solutions, including real‑time streaming pipelines
applying strong data modeling principles such as partitioning strategies, schema evolution handling, and metadata management
building and maintaining CI/CD pipelines using GitLab or Jenkins
automating infrastructure provisioning with Terraform
ensuring code quality through unit testing, code reviews, and adherence to engineering best practices
Requirements
8+ years of strong hands‑on experience in Python, SQL, and PySpark
Solid understanding of distributed data processing frameworks and cloud‑native architectures
Proven experience working within the AWS data ecosystem
Experience delivering solutions in large‑scale, fast‑paced environments with a strong focus on automation and data quality
Good working knowledge of Data Warehousing (DWH) concepts
Data Lakes / Lakehouse architectures
Ability to design and support production‑grade data pipelines end‑to‑end
Basic hands‑on experience with Terraform and CI/CD pipelines
Familiarity with Delta Lake / Iceberg / Hudi
Exposure to real‑time streaming architectures
AWS Data Engineer or Advanced‑level AWS certifications (preferred)
Benefits
flexible, supportive environment
well-being prioritized
personalized development goals aligned with your ambitions
Google Data Architect creating and optimizing data solutions using GCP technologies. Collaborating with teams to enhance enterprise data architecture across business functions.
Azure Lead Data Engineer designing and developing ETL/ELT pipelines with Azure Data Factory and Snowflake. Collaborating with cross - functional teams in a cloud - native environment.
Principal Data Engineer leading Azure platform designs and implementations for enterprise solutions at UBDS Group. Mentoring teams and driving high engineering standards in hybrid environments.
Data Engineer designing and maintaining the data systems for Skiffra’s AI - native orchestration platform. Collaborating closely with product and engineering teams for data integration and system design.
Data Architect responsible for the integrity and reliability of Patient Services data in Life Sciences. Ensuring analytics - ready data through strategic vendor collaboration and data stewardship.
Project & Data Engineer providing operational support and data management for utility service projects in the Greater Los Angeles area. Involves invoice processing, data accuracy, and system coordination.
Senior Data Engineer developing scalable data architectures and integrating data ecosystems at Porto Bank. Ensuring data quality and effective pipeline development for various business teams.
Data Engineering Advisor designing data flow management systems to support advanced analytics at Desjardins Group. Collaborate with teams to enhance data value and transformation.
Founding Staff Data Engineer building and leading data engineering team for AI - driven art valuation platform. Establishing architecture and standards for data systems and pipelines.