Ensure adherence to data privacy, security, observability, and governance standards
Collaborate with architects and senior engineers to design resilient, high-performance data solutions
Build scalable data pipelines to ingest, process, and expose data using Snowflake, AWS Redshift, Kinesis, Lambda, DynamoDB, S3, Glue, Apache Storm, and others
Develop batch and streaming data solutions using Apache Kafka, Apache Spark, Airflow, and Delta Lake
Automate ingestion from external sources, ensuring reliability and data quality
Design and implement event-driven data architectures with Kafka Streams and Kinesis Data Firehose
Implement Data Quality, Data Lineage, and Observability frameworks
Collaborate with Site Reliability Engineers to ensure stability and uptime of production pipelines and services
Work with Data Scientists to ensure high-quality data for Machine Learning models and integrate them into production-ready pipelines
Stay up-to-date and continuously upskill on emerging technologies, frameworks, and best practices in data engineering and cloud platforms
Requirements
Undergraduate degree in Computer Science, Engineering, Data Science or related discipline
3+ years of hands-on experience in data profiling, ETL development, SQL optimization, and testing
Proven experience delivering large-scale data solutions with AWS technologies (Kinesis, Athena, Redshift, DynamoDB, Lambda, S3)
Proficiency with modern data platforms and tools, including Snowflake, Databricks, and Apache Storm
3+ years of experience working with SQL and NoSQL databases, particularly MongoDB
Experience developing batch and streaming data solutions using Apache Kafka, Apache Spark, Airflow, and Delta Lake
Experience with streaming frameworks and event-driven architectures (Kafka Streams, Kinesis Data Firehose)
Experience automating data ingestion, ensuring reliability and data quality
Familiarity with CI/CD tools and practices like Ansible, CloudFormation, and Jenkins
Experience implementing Data Quality, Data Lineage, and Observability frameworks
Experience collaborating with Site Reliability Engineers to ensure production stability and uptime
Experience working with Data Scientists to productionize Machine Learning models
Excellent interpersonal and communication skills
Desire to stay up-to-date and continuously upskill on emerging data engineering and cloud technologies
Data Engineer developing architecture and pipelines for data analytics at NinjaTrader. Empowering analysts and improving business workflows through data - driven solutions.
Data Engineer joining Alterric to collaborate on data platform projects and analytics solutions. Working with Azure Cloud technologies to ensure data quality and integrity for informed decision - making.
Data Engineer at Kyndryl transforming raw data into actionable insights using ELK Stack. Responsible for developing, implementing, and maintaining data pipelines and processing workflows.
Senior Data Engineer at Clorox developing cloud - based data solutions. Leading data engineering projects and collaborating with business stakeholders to optimize data flows.
Data Engineer building solutions on AWS for high - performance data processing. Leading initiatives in data architecture and analytics for operational support.
Senior Data Engineer overseeing Databricks platform integrity, optimizing data practices for efficient usage. Leading teams on compliance while mentoring a junior Data Engineer.
Associate Data Engineer contributing to software applications development and maintenance using Python. Collaborating with teams for clean coding and debugging practices in Pune, India.
Data Engineer focusing on development and optimization of data pipelines in an insurance context. Ensuring data integrity and supporting data - driven decision - making processes.
Lead Data Engineer responsible for delivering scalable cloud - based data solutions and managing cross - functional teams. Collaborating with global stakeholders and ensuring high - quality project execution in a fast - paced environment.