AzureCloudInformaticaPySparkPythonSQLUnityVaultMLData EngineeringData WarehousingAnalyticsDatabricksSaaSCommunicationRemote Work
About this role
Role Overview
Define the enterprise data architecture: Own the conceptual, logical, and physical data models for MeridianLink's analytical and operational data platform, including source-aligned, integrated, and consumption-ready layers.
Build the meta-model: Design and maintain a meta-model that captures entities, relationships, business definitions, ownership, lineage, sensitivity classifications, and SLAs — and make sure it is wired into our tooling, not stuck in a slide deck.
Drive the lakehouse strategy: Architect our medallion (bronze / silver / gold) Delta Lake patterns on Databricks; define standards for partitioning, clustering, schema evolution, slowly changing dimensions, and historical reproducibility.
Be hands-on: Write PySpark, SQL, and Delta Lake code. Build reference implementations, prototype patterns, review pull requests, and personally model critical domains rather than delegating every detail.
Lead data integration design: Set patterns for ingestion through Informatica Data Management Cloud (IDMC) and direct Databricks pipelines, including CDC, batch, streaming, and API-based sourcing from our SaaS products and third-party systems.
Champion data governance and lineage: Partner with data governance, security, and compliance leaders to operationalize cataloging, lineage, classification, masking, and access controls across the platform (Unity Catalog, IDMC, and adjacent tools).
Standardize data modeling practices: Establish the standards, naming conventions, and review processes used by the Data Engineering team. Coach engineers on dimensional modeling, Data Vault, and other techniques where they best fit the use case.
Partner across the business: Work closely with Product, Engineering, Analytics, ML, Finance, Risk, and Customer-facing teams to translate business needs into durable data designs.
Influence the roadmap: Identify gaps in tooling, capability, and skill; propose investments; and drive multi-quarter initiatives that materially improve how MeridianLink uses its data.
Requirements
12–15+ years of progressive experience in data engineering, data warehousing, and data architecture roles, with at least the most recent several years at the architect level.
Demonstrated experience as a Data Architect at a SaaS company in the FinTech or financial services software space (lending, banking, payments, capital markets, insurance, or a closely related domain).
Deep, hands-on expertise with Databricks and PySpark on Azure, including Delta Lake, Unity Catalog, structured streaming, and performance tuning at scale.
Production experience with Informatica Data Management Cloud (IDMC) — or comparable enterprise integration platforms — for ingestion, transformation, and metadata-driven pipelines.
Proven track record of designing and implementing detailed meta-models and end-to-end data models (conceptual, logical, and physical) that have shipped to production and stood up over time.
Strong command of dimensional modeling (Kimball), Data Vault 2.0, and modern lakehouse patterns, including the ability to choose the right approach for the right use case.
Expert SQL skills and strong proficiency in Python/PySpark; comfortable writing the code, not just the diagrams.
Demonstrated experience implementing data governance, lineage, and metadata management programs (e.g., Unity Catalog, IDMC Data Governance, Collibra, Atlan, or similar).
Working knowledge of FinTech-relevant regulatory and compliance considerations (e.g., GLBA, SOC 2, PCI, NIST, state lending regulations) and how they shape data design.
Excellent written and verbal communication skills; able to explain complex data concepts to engineers, executives, customers, and auditors.
Tech Stack
Azure
Cloud
Informatica
PySpark
Python
SQL
Unity
Vault
Benefits
Insurance coverage (medical, dental, vision, life, and disability)