OhioHealth is more than a health system; it is a belief system that emphasizes wellness and the importance of expert guidance in health. They are seeking an experienced Lead Data Engineer to oversee the architecture, development, and maintenance of their data platform, ensuring data quality and efficiency while collaborating with various teams to deliver complex applications and services.
Responsibilities:
- Collaboration - Works with architects, Business Analysts, Product Owners, and peers providing technical expertise by enforcing the vision of the application architecture and safeguarding the integrity of the application environment. Reviews and provides input into architectural designs and requirements
- Functional Subject Matter Expertise - Functions as a subject matter expert and advocate for the software applications supported
- Technical Expertise - Serves as the deep technical expert for the applications supported by team
- Support cross team technical collaboration
- Identify risks and design contingency plans
- Analyze existing operations and evaluate options for improvements
- Conduct security audits to identify areas of improvement
- Guides their teams through technical issues and challenges
- Assists with designing the governance, quality assurance and standards activities associated with ensuring technology architecture compliance of projects and products
- Leads technology change to deliver business outcomes while also virtually leveraging subject matter experts from any other needed technical area
- Oversees, or consults on, technology implementation and modification activities (for example, projects and products), particularly for new or shared solutions
- Creates and delivers working, production-ready IT services that can be easily leveraged by agile teams
- Adaptability to new and emerging technology, oversees PoCs
- Provides the detailed designs, reviews the implementation by the developers, ensures quality, and best practices such as automated unit tests, integration tests, continuous integration
- Technical Documentation - Creates, reviews, and approves technical documentation for applications supported
- Technology Strategy - Leads the development of technology strategy and roadmap within a given area of specialty including design, documentation, communication, and implementation of new or existing standards for infrastructure technologies while remaining current on industry trends
- Vendor Management - Works with 3rd party vendors to validate technical aspects and assess feasibility concerns and develop reporting metrics, dashboards, and status reporting to drive outcomes and meet expectations for quality and performance
- Operations - Works with internal customers and 3rd party vendors to communicate, coordinate, manage expectations, approve, and assist with operational processes such as crisis management, support escalation, change management, release management, capacity management, and other ITIL processes
- Research and Analysis - Analyzes, plans, justifies, schedules, and coordinates the implementation and maintenance of the specialty area’s administrative, operational, and financial initiatives including TCO, optimization, and efficiency
Requirements:
- Bachelor's Degree: Computer and Information Science
- Bachelor's or Master's degree in Computer Science, Information, or a related field
- Minimum of 12 – 15 years of experience in data engineering, with a focus on ETL and Cloud technologies (can be less than 10 years)
- Proficiency in Informatica ETL tools, SQL, and Azure Cloud platforms (preferably Azure but others acceptable)
- Experience in Informatica IDMC - Architect, Design and develop ETL processes on IDMC and Informatica suite of tools. Familiarity with MDM Architecture and Data Flow
- Experience in potential migration from existing data platforms to Databricks, Microsoft Fabric
- Must have working experience with Azure based Data Pipelining, Scheduling and Monitoring and pyspark with ability to debug troublesome pipelines. Must have hands on expertise dealing with data pipelines
- Strong working experience with Big Data technologies (Spark, Data Bricks) for Data integration, & processing (ingestion, transformation, curation, etc), preferably on Azure cloud, and a clear understanding of how the resources work and integrate with cloud and on-prem
- High level of proficiency with database and data warehouse development, including replication, staging, ETL, stored procedures, partitioning, change data capture, triggers, scheduling tools, cubes, and datamarts
- Experience working with backend languages such as Python
- Strong computer literacy and proficiency in data manipulation using Analytics tools/platform like Databricks, Azure Fabric using Spark engine
- Expertise in at least one technology stack designing, developing, testing, and/or delivering complex software (i.e., Java, Python, Pyspark)
- Excellent debugging, troubleshooting, and analytical skills
- Strong analytical and problem-solving skills with the ability to own, troubleshoot and resolve complex data issues
- Collaborate with Architects and Managers to develop Metrics & KPI's
- Identify any technical risks and forming contingency plans as soon as possible
- Good communication and collaboration skills to work effectively with cross-functional teams
- Experience working in an Agile development environment preferred
- Experience building data pipelines to support ML workflows a plus
- Experience working with geographically distributed teams (different time zones)