Senior Data Engineer curating and scaling data pipelines for healthcare tech company Arbital Health. Building AI/ML data solutions and collaborating with cross-functional teams.
Responsibilities
Leverage your advanced skills in Python and SQL and with platforms like Databricks and AWS to build highly scalable data pipelines and warehouses
Design and maintain context engineering pipelines (embedding generation, indexing, vector storage) for conversational AI, collaborating with the Senior AI Engineers and Data Scientists.
Develop and deploy scalable AI/ML data pipelines, specializing in data prep/serving for LLMs and RAG systems.
Stay ahead of emerging tech and integrate it into business solutions
Automate data workflows for ingestion, cleansing, quality assurance, enrichment, and aggregation
Ensure data accuracy, integrity, privacy, security, and compliance through automated quality control procedures
Collaborate with an actuarial science team
Contribute to the design, implementation, and overall development of our products
Drive innovation and deliver valuable features for our customers
Requirements
Experienced in data-intensive, full-stack development projects
Experience with data/AI architecture and MLOps
Proven, hands-on experience designing and implementing data-intensive solutions using LLMs, vector databases, embeddings, and context engineering techniques (RAG, summarization).
Able to work entrepreneurially – self-motivated, ambitious, and fast-paced
Able to ship extremely high caliber code and build exceptional products
High level of attention to detail
Ability to perform under minimal supervision with accountability for specific objectives and work in a rapidly changing, ambiguous start-up environment
Passionate about improving and innovating
Startup experience is highly preferred
Our team works hybrid from the San Francisco Bay Area. We will prioritize candidates who are able to work 1-2x per week from our office and we will consider highly qualified remote candidates who are able to travel for in-person collaboration in San Francisco at least one full week per month.
Benefits
Generous equity grants of ISO stock options
We offer an exceptional benefits package with high employer-paid contributions for health, dental, and vision insurance
4% 401(k) match
Flexible PTO, a weeklong winter shutdown, and 10 holidays each year
Occasional travel required - Quarterly team offsites
Senior AI & Data Engineer developing and implementing AI solutions in collaboration with clients and teams. Working on projects involving generative AI, predictive analytics, and data mastery.
Consultant driving IA business growth in Deloitte's Artificial Intelligence & Data team. Delivering innovative solutions using data analytics and automation technologies.
Data Engineer responsible for managing data architecture and pipelines at Snappi, a neobank. Collaborating with teams to enable data processing and analysis in innovative banking solutions.
Data Engineer at Destinus developing the data platform to support production and analytics needs. Involves migrating Excel sources to Lakehouse and integrating ERP systems in a hybrid role.
Senior Data Engineer developing solutions within the Global Specialty portfolio at an insurance company. Engaging with diverse business partners to ensure high quality data reporting.
Data Engineer at UBDS Group focusing on designing and optimizing modern data platforms. Collaborating in a multidisciplinary team to develop reliable data assets for analytics and operational use cases.
Data Engineer (dbt) at SDG Group involved in all phases of data projects. Collaborate on data ingestion, transformation, and visualization in a hybrid environment.
Data Consultant at SDG Group specializing in Data & Analytics projects. Collaborate on technical - functional definitions, ETL, data modeling, and visualization for cloud solutions.
Senior Data Engineer responsible for growing customer - defined targeting calculations and developing key/value databases for real - time data processing.
Data Engineer developing and maintaining the Data Lakehouse platform using Microsoft Azure technology stack at RBC. Collaborating with business and technology teams to enhance data ingestion and modeling processes.