Senior Data Engineer joining Financial Crime team to build data pipelines for fraud detection. Working with complex datasets in Databricks and collaborating with cross-functional teams.
Responsibilities
Build and optimize data ingestion pipelines using Python and PySpark to collect and transform data from multiple sources (transactions, KYC, AML, authentication, devices, logs, etc.).
**Proficiency in SQL (PostGres preferred) **
**Design and maintain data model that support Financial Crime/Fraud detection, profiling, and entity resolution. **
**Implement data quality checks and ensure data reliability across environments.
**Collaborate closely with Data Scientists, Analysts, Compliance, Operations and our Product/Feature teams to operationalize models and rules. **
Utilize jobs, workflows, APIs and streaming to manage large-scale data processing workloads.
Integrate with external systems (e.g. sanctions, ID&V, biometrics and authentication systems) to enrich risk and identity data.
Support **automation and monitoring** of ETL processes to improve operational efficiency.
Requirements
Bachelor’s degree.
**5+ years of experience **
**Strong skills in Python, PySpark, Scala and Advanced SQL (preferably PostGres) **
**Hands-on experience with Databricks, Snowflake, Fabric or similar **
**Experience working with structured and unstructured data in a production environment. **
**Experience with Agentic AI, MLFlow, ML models, Eval **
**Secure Coding practices – testing/QA **
**Comfortable with cloud-based data platforms (preferably AWS). **
**Good communication skills in English — able to collaborate with cross-functional teams in an international environment. **
**Proficiency in working with Text, Delta, Parquet, JSON, CSV, and XML data formats. **
**Working knowledge of Spark structured streaming. **
**AWS infrastructure experience, specifically working with S3. **
**Solid understanding of git-based version control, DevOps, and CI/CD. **
**Experience of working on Atlassian stack a plus. **
**Knowledge of common web API frameworks and web services. **
Strong teamwork, relationship, and client management skills, and the ability to influence peers and senior management to accomplish team goals.
Willingness to embrace modern technology, best practice, and ways of work.
**Nice to Have: **
Experience in **Financial Crime/AML, KYC, **or** fraud detection** systems.
Familiarity with **Entity Resolution frameworks** (e.g., Quantexa, Sensing, open source Entity Resolution tools).
Experience with **data streaming frameworks** (Kafka, Spark Streaming, MQ).
Benefits
Be part of a **mission-driven** team tackling real-world financial crime problems.
Work with **modern data tech stack** with Agentic AI and advanced ML.
**Hybrid working model **with flexible hours.
International and collaborative culture — working with colleagues across **Vietnam, Singapore, Philippines and South Africa**.
Competitive salary, performance bonuses, and learning support.
Medior Data Engineer handling Azure migrations for a major urban mobility client. Focused on data pipeline development and ensuring platform reliability with cutting - edge technologies.
Developing ML and computer vision solutions for cutting - edge autonomous vehicle dataset pipeline at Mobileye. Collaborating across teams for data curation and advanced perception algorithms.
Data Migration Lead in a hybrid role managing data migration for a major transformation programme in the media sector. Collaborating with various teams to ensure data integrity and successful migration.
Consultant ML & DataOps at Smile integrating data science projects for major clients. Designing MLOps solutions and enhancing data governance in a collaborative environment.
Data Engineer developing and maintaining data pipelines for Coolbet’s analytical services. Working within an Agile framework to ensure data reliability and efficiency.
API Data Engineer developing innovative data - driven solutions and advancing data architecture for AI Control Tower. Building and integrating APIs and data pipelines to support organizational needs.
Journeyman Data Architect supporting Leidos' enterprise data and analytics program for the Department of War. Collaborating on solutions for data architecture, cloud environments, and governance.
Senior Software Engineer developing backend services and data infrastructure for integrated products at Booz Allen. Collaborating with a small elite team to deliver reliable and scalable services.
AWS Streaming Data Engineer developing software and systems in a fast, agile environment. Utilizing experience with real - time data ingestion and processing systems across distributed environments.
Mid - level Data Engineer ensuring efficient data transformation and integration for data annotation projects. Collaborating with teams to optimize data quality and performance in pipeline operations.