Staff Data Engineer designing automated data infrastructure for early cancer detection technologies at GRAIL. Collaborating with engineers and scientists to streamline operations and ensure data quality.
Responsibilities
Be a part of a highly collaborative team focused on delivering value to cross-functional partners by designing, deploying, and automating secure, efficient, and scalable data infrastructure and tools—reducing manual efforts and streamlining operations.
Help model Grail data to ensure it follows FAIR principles (Findable, Accessible, Interoperable, and Reusable).
Drive the design, deployment, and automated delivery of data infrastructure, standardized data models, datasets, and tools.
Integrate automated testing and release processes to improve the quality and velocity of software and data deliveries.
Collaborate with cross-functional teams—ranging from Research to Clinical Lab Operations to Software Engineering—to provide comprehensive data solutions from conception to delivery.
Ensure all software and data meet high standards for quality, clinical compliance, and privacy.
Mentor fellow engineers and scientists, promoting best practices in software and data engineering.
Requirements
B.S. / M.S. in a quantitative field (e.g., Computer Science, Engineering, Mathematics, Physics, Computational Biology) with 8+ years of related industry experience, or Ph.D. with at least 5 years of related industry experience or equivalent.
Extensive experience developing with relational databases, data modeling principles, data pipeline tools, and workflow engines (e.g., SQL, DBT, Apache Airflow, AWS Glue, Spark).
Extensive experience with DevOps practices, including CI/CD pipelines, containerized deployment (e.g., Kubernetes), and infrastructure-as-code (e.g., Terraform).
Experience supporting data science and machine learning data pipelines, preferably in the context of biological data analysis and using Python.
Ability to embrace uncertainty, navigate ambiguity, and collaborate with product teams and stakeholders to refine requirements and drive toward clear engineering objectives and designs.
You are a strong written and verbal communicator and can adapt your communication style and the level of detail to your audience.
Senior Data Engineer / Data Architect at Node.Digital focusing on designing data architectures and managing pipelines. Collaborating across teams to support enterprise application delivery.
Data Engineering Lead responsible for data pipeline design and optimization at Mars. Leading a talented team to drive impactful data solutions across North America.
Data Engineering Developer intern participating in secure data flow creation at Intact. Collaborating on data engineering using Python and cloud technologies for an enterprise data platform.
Data Engineer responsible for building and maintaining data transformation pipelines at OnePay. Collaborating across teams in a mission - driven fintech environment.
Senior Developer within Enterprise Data Management at LPL Financial. Responsible for supporting data management projects and collaborating with business partners and developers.
Lead Data Engineer at Capital One solving complex business problems with data and emerging technologies. Collaborating across Agile teams to deliver cloud - based technical solutions.
Join Luminor as a Mid/Senior Data Engineer focusing on data engineering within risk and finance reporting. Design and optimize data systems supporting evolving regulatory requirements in a dynamic banking environment.
Join Luminor as a Mid/Senior Data Engineer focusing on data engineering within risk and finance reporting. Collaborate to design scalable data architectures and support regulatory requirements while enhancing data integration processes.
Senior GCP Data Engineer designing, building, and optimizing data platforms on GCP. Collaborating with product teams to deliver high - performance data solutions.
Snowflake Data Engineer optimizing data pipelines using Snowflake for a global life science company. Collaborate with cross - functional teams for data solutions and performance improvements in Madrid.