Data Engineer to build and support data pipelines and models for analytics and machine learning. Working with ADI's Product Data Science team to leverage extensive data and automate processes.
Responsibilities
Designing and implementing data pipelines to efficiently extract, transform, and load (ETL) from various data sources into a central repository or data warehouse
Deploy and maintain scalable production pipelines for AI models, ensuring seamless integration with existing systems, continuous performance monitoring, and iterative improvements based on real-world feedback
Working on transformations and contributing to building the dimensional model of ADI product data
Building and maintaining scalable and robust data infrastructure, including data lakes, and distributed computing systems that are equipped to handle large volumes of data
Implement, and build data solutions using Spark, Python, Databricks, and the AWS ecosystem (S3, Redshift, EMR, Athena, Glue).
Implementing data validation and cleaning processes to ensure data quality and monitoring data pipelines for errors or anomalies
Optimizing data storage and retrieval processes to enhance performance and reduce latency, particularly through techniques such as indexing, partitioning, and caching
Implementing data security and privacy measures, including data encryption, access controls, and compliance with data governance policies and regulations
Collaborating with data scientists and data analysts to understand metrics and data requirements and to provide them with access to the necessary data sets and data systems
Adapt, learn, grow, and teach to deliver world-class products and platforms
Bring a curious mind to current business problems and opportunities, expanding your understanding of all products in the ADI ecosystem
Requirements
Strong programming skills and in Python
Strong expertise in database and query languages like SQL, NoSQL
Knowledge of data engineer skills such as data management, data visualization, and familiarity with data architecture
Knowledge and hands-on experience in building scalable data platforms and reliable data pipelines using technologies such as Spark, Databricks, AWS Kinesis, and/or Kafka
Solid understanding of MLOps principles and model versioning
Experience working with large volumes of metadata and schemas
Hands-on experience in ETL/ELT and data integration
Data warehousing knowledge, including data modeling, data security, and data governance understanding
Understanding how the role complements others working with machine learning, data science, algorithms, business intelligence
Benefits
Health package
Insurance in case of serious illness, surgical intervention, professional illness, and insurance from the consequences of an accident
Flexible working hours
English classes during working hours
Employee referral bonus program
Corporate social events and team buildings
Food and drinks: Free use of coffee machines, free fruit and snacks
Data Management professional at Kyndryl involved in creating innovative data solutions and ensuring the seamless operation of complex data systems. Collaborating with teams to transform requirements into scalable database solutions.
Software Engineer designing and developing scalable data processing applications on cloud infrastructure for Thomson Reuters. Collaborating with Data Analysts on AI - enabled solutions for data management and insight generation.
Manager of Data Platform overseeing AWS cloud infrastructure and Snowflake data warehouses for Thomson Reuters. Leading the design and implementation of data processing applications in a hybrid role located in Bengaluru.
Senior Data Engineer designing scalable data pipelines and solutions for Enterprise Data Lake at Thomson Reuters. Collaborating across teams to ensure efficient data ingestion and accessibility.
Senior Data Engineer at Technis developing scalable data pipelines and solutions for innovative connected spaces products. Collaborating within a cross - functional team to deliver high - quality data - driven outcomes.
Data Architect designing and implementing data architectures supporting analytics and ML for federal clients. Collaborating with teams to translate mission needs into robust data solutions.
IT Data Engineer developing data pipelines and integrations for Scanfil Group's global IT organization. Collaborating across teams to enhance data solutions and reporting capabilities.
Data Engineer developing Azure data solutions at PwC New Zealand. Responsibilities include data quality monitoring, pipeline development, and collaboration with stakeholders in a supportive environment.
Senior Data Engineer designing and implementing the Enterprise Data Platform at Stellix. Focusing on analytics and insights with a growth path to Principal Data Engineer or Data Architect.