Qualified Health is seeking a Data Engineer to support partner integration in transforming healthcare. The role involves building data pipelines that power AI systems used by clinicians across major U.S. health systems, directly impacting patient care.
Responsibilities:
- Build and maintain data ingestion pipelines that pull from Epic Clarity databases, FHIR APIs, and HL7 feeds into our Databricks lakehouse architecture — the technical backbone of every AI workflow we deliver
- Execute data quality validation and mapping verification for each new workflow deployment — ensuring AI models receive clean, accurate, clinically relevant data. Your QC catches the problems before they reach production
- Support 2-4 health system partners as part of a dedicated integration pod, working closely with your Director and directly with customer data teams
- Develop and optimize SQL queries against complex healthcare data models (hundreds of tables, billions of rows) including clinical, billing, and operational data
- Contribute to defining reusable data transformation patterns that accelerate future partner onboarding — the patterns you build today are used by every pod tomorrow
Requirements:
- 4+ years as a data engineer building production ETL/ELT pipelines — you've shipped real systems, not just notebooks
- Strong SQL skills (complex joins, window functions, CTEs across large healthcare datasets). Healthcare SQL is its own discipline — you either know it or you're ready to master it
- Proficiency with Python and PySpark for data transformation at scale
- Experience with cloud data platforms (Databricks preferred; Snowflake, BigQuery, or Redshift accepted — you'll learn Databricks fast if you haven't already)
- Familiarity with FHIR data models or willingness to learn healthcare data standards quickly — we'll invest in your ramp
- Azure cloud services experience (ADLS Gen2, Key Vault, networking basics)
- Bachelor's degree in Computer Science, Engineering, Data Science, Mathematics, or related technical field
- Epic Clarity/Caboodle experience or healthcare data background — this is a huge accelerant
- Delta Lake, Delta Sharing, or lakehouse architecture experience
- HL7 message processing or real-time healthcare data pipelines
- Experience working in a multi-tenant SaaS environment