Data Engineer joining Contour Software's development team to handle various big data projects. Focus on cloud data lake architecture and machine learning implementation.
Responsibilities
Write efficient queries to extract and analyze large datasets while ensuring seamless systems integration by developing robust data workflows.
Design, develop, and maintain scalable ETL solutions for data pipelines, enhancing and optimizing existing processes to meet evolving business needs.
Develop and expand the organization’s data technology stack to support advanced data processing and analytics needs, leveraging tools like AWS MapReduce and PySpark for big data handling.
Apply machine learning techniques to clean and process data, supporting future initiatives to integrate machine learning into analytics workflows.
Continuously evaluate and adopt the most efficient tools and technologies for data extraction, transformation, analysis, and integration.
Build a strong understanding of customer business needs to deliver tailored and impactful data solutions, collaborating with stakeholders to translate business requirements into technical implementations.
Utilize AWS services, including MapReduce and related tools, to enable scalable data processing and pipeline optimization while ensuring security, scalability, and cost-efficiency.
Monitor and improve the performance of data systems and processes, staying updated on industry trends and emerging technologies in data engineering and analytics.
Requirements
Minimum 3 years of experience with AWS cloud data lake architectures including services like S3, Glue, Athena, and Redshift.
Hands on experience with Apache Airflow for designing and managing complex data workflows.
Deep understanding of data warehouse concepts, architectures, and structures.
Expertise in cloud technologies, particularly AWS services, including AWS Glue, AWS Data Lake, and Amazon EMR for big data processing.
Proven ability to design and build robust, scalable data pipelines for big data processing and transformation.
Hands-on experience with big data tools and frameworks such as PySpark, Databricks, and related technologies.
Proficiency in programming languages such as Python for data manipulation, transformation, and analysis.
Advanced Python development skills, including experience with AWS SDKs (e.g. boto3) for interacting with cloud services.
Proficiency in Apache Kafka for real-time data streaming and event-driven architectures.
Strong SQL skills with a focus on complex query development and optimization.
Familiarity with NoSQL databases such as DynamoDB or MongoDB.
Experience integrating Python applications with RESTful APIs and external services.
Deep understanding of data security best practices, including encryption at rest and in transit.
Hands-on experience in implementing AWS KMS (Key Management Services) for managing encryption keys.
Familiarity with IAM policies, VPC configurations, and security groups for securing data pipelines.
Familiarity with machine learning concepts and their application to data cleaning and integration tasks.
Advanced English proficiency, with excellent communication skills for collaborating with stakeholders.
Data Platform Expert developing and maintaining data solutions for analysis and reporting at Magna Electronics. Collaborating with various teams to enhance data - driven decision making and insights.
Data Engineer at Mobiz designing, building, and maintaining scalable data solutions for analytics. Collaborating with teams to leverage modern cloud technologies and improve data - driven decision - making.
Head of Data Engineering at Envitia overseeing data architecture services for public sector programs. Leading service mobilization with client stakeholders in a hybrid work environment.
Applied AI Health Data Architect - Senior Manager at PwC designing data architecture for healthcare operations. Contributing to innovative data solutions and mentoring teams for operational excellence.
CRM Functional Data Migration Engineer at Data - Core Systems, Inc. managing Dynamics 365 CE solutions and delivering technical expertise in a consulting role.
Data Engineer role for Alberta Securities Commission, focusing on data pipelines and infrastructure. Involves designing, building, and securing data assets for organizational decision - making.
Data Architect supporting UK MOD programme delivering data and integration capabilities for Defence. Join Amentum's growing IT team focused on critical national infrastructure and security projects.
Enterprise Architect Sr position at PNC collaborating with technology stakeholders to build enterprise data architecture. Focus on data models, governance, and guiding data decisions for business strategy.
Data Engineer architecting and maintaining data products using Databricks and Power BI for Nitro's digital transformation efforts. Collaborating to optimize processes and leverage cloud technologies.
Lead Data Engineer driving major transformation in enterprise technology at Capital One. Collaborating with Agile teams to develop, test, and implement innovative full - stack solutions.