McKesson is an impact-driven, Fortune 10 company that touches virtually every aspect of healthcare. They are seeking a highly skilled Senior Data Engineer specializing in claims data to design, develop, and maintain robust data pipelines and infrastructure to support advanced analytics and reporting initiatives related to healthcare claims.
Responsibilities:
- Design, develop, and optimize scalable and reliable data pipelines for extracting, transforming, and loading (ETL) large volumes of healthcare claims data from various sources into our Data Platform and other analytical platforms
- Work closely with data scientists, analysts, and product managers to understand data requirements and translate them into efficient data models and technical solutions
- Implement and maintain data quality checks, data validation processes, and data governance standards to ensure the accuracy, completeness, and consistency of claims data
- Develop and manage data schemas, data dictionaries, and documentation for claims-related data assets
- Monitor and troubleshoot data pipeline performance, identify bottlenecks, and implement optimizations to improve efficiency and reduce processing times
- Collaborate with engineering teams to integrate data solutions with existing and new applications and platforms
- Evaluate and recommend new technologies and tools to enhance our data engineering capabilities and improve claims data management
- Participate in code reviews, mentor junior data engineers, and contribute to best practices for data engineering within Ontada
- Ensure compliance with healthcare data regulations (e.g., HIPAA) and internal security policies
Requirements:
- Degree or equivalent and typically requires 7+ years of relevant experience
- Bachelor's degree in Computer Science, Engineering, Information Systems, or a related quantitative field
- 7+ years of professional experience in data engineering, with a significant focus on healthcare claims data
- Expert proficiency in SQL and experience with various relational and non-relational databases (e.g., Databricks, Postgres)
- Strong programming skills in Python, with experience in data manipulation libraries (e.g., Pandas, NumPy) and scripting for data processing
- Extensive experience with ETL/ELT tools and platforms (e.g.Azure Databricks, dbt, Apache Spark, Azure Data Factory, Snowflake)
- Demonstrated experience with cloud data platforms (Azure preferred), particularly with services like Azure Data Factory, AKS, GitHub, GitHub Actions
- Solid understanding of healthcare claims data standards (e.g., X12 EDI 837, 835) and common claims data elements
- Experience with data warehousing concepts, dimensional modeling, and data lake architectures
- Knowledge of data governance principles, data security best practices, and compliance requirements in healthcare
- Excellent problem-solving skills, attention to detail, and ability to work independently and collaboratively in a fast-paced environment
- Strong communication skills, both written and verbal, with the ability to explain complex technical concepts to non-technical stakeholders
- Travel up to 10%
- Master's degree preferred