Senior Data Engineer designing scalable data pipelines and solutions for Enterprise Data Lake at Thomson Reuters. Collaborating across teams to ensure efficient data ingestion and accessibility.
Responsibilities
Design and implement end-to-end solutions to ingest data from multiple structured, semi-structured, and unstructured sources into the Enterprise Data Lake
Conduct source data analysis in collaboration with data subject matter experts to build robust data pipelines and thoroughly document data transformations
Develop and maintain scalable, sustainable ETL/ELT pipelines for both on-premises and cloud environments, ensuring efficiency, reliability, and performance
Innovate with new approaches for source ingestion, ELT data pipelines
Design and build data products that serve downstream consumers including data science, business analytics, and reporting teams
Perform end-to-end deployment of data pipelines using CI/CD processes
Collaborate with business analysts, data architects, and cross-functional teams to gather requirements, define data contracts, and deliver solutions aligned with business goals
Effectively communicate and liaise with data management teams embedded across the organization
Manage and prioritize technical debt while maintaining a high standard of code quality, pipeline reliability, and system scalability
Evaluate and adopt AI-powered tools and frameworks to improve pipeline development, data quality monitoring, and operational efficiency
Requirements
7+ years of demonstrated expertise in data engineering
Senior Data Engineer: AWS, Python, Spark, Snowflake, SQL, CI/CD, data modeling, AI
Core Data Engineering Expert in Data Pipeline Development: ETL/ELT tooling for on-premises and cloud environments
Expertise in ETL tools, preferably Informatica
Working experience building sustainable data pipelines to Data Lakes, Data Warehouses, and Data Marts
Expertise in handling semi-structured and unstructured data
Big Data experience with distributed processing frameworks
Highly proficient in SQL
Proficient in Python and/or Spark for data engineering workloads
Highly experienced in working with relational and non-relational databases, including Snowflake
Strong knowledge of ingesting data from APIs, RDBMS, SaaS Application (CRM and ERP), Social Media and Realtime/Event streaming sources
Hands-on experience on DBT Transformation, Shell Scripting, Store procedures etc
Proficient in AWS services including S3, EC2, Lambda, IAM, VPC, Load Balancer, Glue, and RDS (Mandatory Hands-on AWS Glue)
End-to-end deployment of data pipelines using CI/CD processes
Experience or working knowledge of Terraform for infrastructure deployment
Strong understanding of data modeling — 3NF and dimensional modeling
Information systems design and solution design experience
Business analysis capabilities to translate requirements into data engineering solutions
Effective stakeholder management across technical and business audiences
Ability to mentor junior engineers, conduct code and design reviews, and foster a culture of continuous learning and improvement
Benefits
Hybrid Work Model: We’ve adopted a flexible hybrid working environment (2-3 days a week in the office depending on the role)
Flex My Way is a set of supportive workplace policies designed to help manage personal and professional responsibilities
This builds upon our flexible work arrangements, including work from anywhere for up to 8 weeks per year
Career Development and Growth: By fostering a culture of continuous learning and skill development
Industry Competitive Benefits: We offer comprehensive benefit plans to include flexible vacation
two company-wide Mental Health Days off
access to the Headspace app
retirement savings
tuition reimbursement
employee incentive programs
resources for mental, physical, and financial wellbeing
Culture: Globally recognized, award-winning reputation for inclusion and belonging, flexibility, work-life balance, and more
Make an impact in your community with our Social Impact Institute
Manager of Data Platform overseeing AWS cloud infrastructure and Snowflake data warehouses for Thomson Reuters. Leading the design and implementation of data processing applications in a hybrid role located in Bengaluru.
Software Engineer designing and developing scalable data processing applications on cloud infrastructure for Thomson Reuters. Collaborating with Data Analysts on AI - enabled solutions for data management and insight generation.
Senior Data Engineer at Technis developing scalable data pipelines and solutions for innovative connected spaces products. Collaborating within a cross - functional team to deliver high - quality data - driven outcomes.
Data Architect designing and implementing data architectures supporting analytics and ML for federal clients. Collaborating with teams to translate mission needs into robust data solutions.
IT Data Engineer developing data pipelines and integrations for Scanfil Group's global IT organization. Collaborating across teams to enhance data solutions and reporting capabilities.
Data Engineer developing Azure data solutions at PwC New Zealand. Responsibilities include data quality monitoring, pipeline development, and collaboration with stakeholders in a supportive environment.
Senior Data Engineer designing and implementing the Enterprise Data Platform at Stellix. Focusing on analytics and insights with a growth path to Principal Data Engineer or Data Architect.
R&D Data Engineer at DXC, transforming complex data into digital assets for global analytics and Smart Lab solutions. Collaborating on ELN and LIMS tools for enhanced data management.
Data Engineer role focusing on data pipelines and processing at 42dot, a mobility AI company. Responsibilities include data collection, schema management, and pipeline monitoring.