Data Engineer crafting data ingestion and transformation processes for an AI healthcare platform. Collaborating with teams to turn complex healthcare data into actionable insights.
Responsibilities
Own data ingestion, transformation, and curation across Bronze, Silver, and Gold layers of our Databricks-based Medallion Architecture.
Manage and optimize data pipelines using Airflow for orchestration and Airbyte (or similar tools) for multi-source ingestion.
Build and maintain connectors and workflows for APIs, EHR/EMR systems (FHIR), resident life, and IoT/monitoring data sources.
Implement batch and streaming pipelines supporting both analytics and near real-time use cases.
Develop and monitor data quality, validation, and profiling frameworks across ingestion points.
Support AI enablement efforts — preparing data for LLM-based insights, population health analytics, and predictive modeling use cases (e.g., fall risk, medication adherence, staffing optimization).
Collaborate closely with data science to enable curated datasets and semantic layers for Superset and AI query interfaces.
Partner with our infrastructure team to maintain infrastructure as code (Terraform) for data services, ensuring scalability and reproducibility.
Partner with security and compliance officers to move towards HIPAA and SOC 2 alignment for all data storage and processing.
Requirements
4–8 years of hands-on data engineering experience.
Strong proficiency with Airflow and Databricks (Spark, Delta Lake, SQL, Python).
Experience building scalable ingestion pipelines with Airbyte, Fivetran, or custom API connectors.
Solid understanding of Azure data ecosystem (Data Lake, Blob Storage, Key Vault, Functions, FHIR Server, etc.).
Experience implementing and maintaining ETL/ELT pipelines in a HIPAA or regulated environment.
Comfort with both SQL and Python for transformations, orchestration, and testing.
Strong grasp of data modeling, schema evolution, and versioned datasets.
Ability to operate independently and deliver results in a small, fast-moving team.
Experience with FHIR and healthcare data structures and interoperability standards.
Familiarity with vector databases (e.g., pgvector, Pinecone) or embedding pipelines for AI/LLM applications.
Experience with GitHub best practices for maintaining and sharing code.
Familiarity with Superset or other analytics tools for internal visualization.
Understanding of security best practices, including encryption, RBAC, and least-privilege design.
Benefits
Christmas Bonus: 30 days, to be paid in December.
Major Medical Expense Insurance: Coverage up to $20,000,000.00 MXN.
Minor Medical Insurance: VRIM membership with special discounts on doctor’s appointments and accident reimbursements.
Dental Insurance: Always smile with confidence!
Life Insurance: (Death and MXN Disability)
Vacation Days: 12 vacation days in accordance with Federal Labor Law, with prior approval from your manager. + Floating Holidays: 3 floating holidays in addition to the 7 official holidays in Mexico.
Intern Data Engineer at Flutter Studios learning to create applications in the gaming industry. Join a cross - functional team to implement a basic application in an Agile environment.
AI/Data Engineer at Comcast developing data pipelines and AI solutions for audit processes. Leading team efforts to ensure data quality and compliance with audit objectives across business units.
Senior Data Engineer involved in AWS Cloud and Big Data solutions for Financial Crime. Join CommBank's team to tackle complex data - centric problems with advanced technologies.
Senior Data Engineer leading and mentoring a team in building scalable data pipelines for digital transformation projects in an international software company.
Senior Data Operations Engineer at iKnowHow designing and implementing scalable data - driven applications. Focus on data pipelines, APIs, and collaboration across teams for project success.
Data/AI Engineer supporting innovations in corporate travel with AWS technologies at HRS Group. Collaborating with data teams to develop AI solutions and maintain data pipelines.
Data Engineer developing data solutions for reporting and analytics at Assurant. Implementing and optimizing Data Warehouse solutions in cloud and on - premise environments with Agile methodology.
AI Engineer developing AI - driven analytics for trading at Deloitte. Focusing on scalable data pipelines and collaboration with traders for actionable insights.
Data Engineer involved in delivering scalable enterprise systems and digital transformation platforms for Qualysoft. Collaborate with global teams across various industries and technologies.