Principal Data Engineer at Trainline shaping robust data foundations for AI and ML-driven products. Collaborate with cross-functional teams to ensure best practices in data engineering and ML.
Responsibilities
Act as a technical authority across multiple teams, setting standards and patterns for data and ML‑adjacent infrastructure
Embed with ML teams to design, build and evolve data platforms supporting AI and ML workloads
Influence technical direction without direct line management responsibility
Partner with Data Engineering teams outside of ML to build a community and share best practices and findings across all areas
Identify systemic issues and proactively drive improvements across the data ecosystem
Look for short term and strategic opportunities to enhance core platforms with new self-serve enablement features for ML and DE
Partnering with MLEs to design data pipelines supporting model training, inference and experimentation
Designing and reviewing architectures for ML‑ready data platforms
Building and optimising data pipelines using SQL, Spark or Ray and Python
Defining best practices for orchestration using Airflow or similar tools
Supporting API‑driven and event‑based data access patterns
Working with AWS infrastructure such as ECS, vector databases and Bedrock APIs
Reviewing designs and code across teams to raise quality and consistency
Coaching engineers through pairing, design reviews and informal mentoring
Collaborating on innovative AI‑powered product features such as the Travel Assistant
Requirements
Extensive experience as a Senior, Staff or Principal Data Engineer operating across teams
Deep expertise in SQL and Python, with strong experience in Spark or similar tooling.
Strong understanding of orchestration tools such as Apache Airflow
Experience designing data platforms for ML and AI workloads
A track record of introducing new technologies and practices, and handling ambiguity and multiple stakeholders.
Hands‑on experience with AWS infrastructure (e.g. ECS, IAM, data storage, compute)
Familiarity with vector databases and modern AI/ML APIs (e.g. Bedrock)
Experience working closely with Machine Learning Engineers in production environments
Strong system design skills and the ability to influence through technical leadership
Migration Engineer responsible for migrating brands to CoreX platform at GiG. Collaborating on data pipelines and integrity with cross - functional teams.
Migration Engineer for GiG’s Migrations Team managing and executing data migrations and ensuring data integrity. Collaborating with various stakeholders on the CoreX platform migrations.
Palantir Data Engineer supporting production data pipelines at a global data and AI company. Monitoring workflows, troubleshooting issues, and ensuring data system reliability.
Senior Azure Data Architect responsible for designing, deploying, and managing scalable cloud infrastructure on Microsoft Azure. Requires over 9 years of experience in cloud architecture and engineering.
Senior Data Engineer designing and building automated data platforms and pipelines for Coody clients. Delivering efficient solutions in a consultant role across exciting industries and companies.
Senior Data Platform Consultant driving analytics solutions on Databricks for Snap's clients. Leading architecture and delivery while shaping Databricks capability growth within the team.
Data Engineer Associate developing and implementing data solutions within PNC's Asset Management Group. Collaborating on technical solutions using PySpark, Hadoop, and SQL for scalable data systems.
Senior Data Engineer at Assembly working on data integration, transformation, and analytics collaboration. Handling cloud services and data quality across data projects with cross - functional teams.
Senior Data Engineer developing scalable data pipelines and collaborating with cross - functional teams at Technis. Technical guidance in a hybrid work environment based in Lausanne, Switzerland.
Data Engineer designing and maintaining data pipelines for CIEE, a philanthropic institution supporting youth development. Collaborating with Data Analysts for data quality and reliability.