TEKsystems is looking for a Data Engineer to lead the design, build, and optimization of their modern data platform on Microsoft Fabric. The role involves architecting and implementing data warehouse schemas, orchestrating data pipelines, and ensuring data quality and security.
Responsibilities:
- Architect & Build on Fabric
- Design and implement Fabric Data Warehouse schemas (star/snowflake, dimensional modeling) and Lakehouse patterns using Delta/Parquet
- Orchestrate end-to-end pipelines with Data Factory (Fabric Pipelines) and Dataflows Gen2—batch and (where relevant) near-real-time
- Develop robust SQL (T‑SQL/DAX‑aware patterns) and Spark notebooks (PySpark) for complex transformations
- Design semantic models that support BI and ML use cases; optimize for cost and performance (partitioning, indexing, caching, concurrency)
- Implement CDC/merge/upsert strategies for evolving datasets
- Implement data validation & testing (unit, schema, anomaly checks), monitor SLAs, and set up alerting & lineage
- Drive data reliability SLOs (freshness, completeness, accuracy)
- Enforce RBAC, row-level/column-level security, masking, and data retention
- Potentially integrate with Microsoft Purview for catalog, lineage, and policies
- Use Git integration in Fabric for version control and CI/CD (multi-environment deploys, templates, release automation)
- Establish IaC patterns where applicable (e.g., ARM/Bicep/Terraform for related Azure resources)
- Partner with Analytics Engineers, Data Scientists, and Power BI Developers to translate requirements into technical designs
- Mentor engineers and champion best practices (coding standards, reviews, documentation)
Requirements:
- Knowledge and possibly hands-on experience with Microsoft Fabric
- Experience with Azure
- Experience with SQL
- Experience with Data
- Experience with PySpark
- Experience with T-SQL
- Experience with Data Warehouse
- Architect & Build on Fabric
- Design and implement Fabric Data Warehouse schemas (star/snowflake, dimensional modeling) and Lakehouse patterns using Delta/Parquet
- Orchestrate end-to-end pipelines with Data Factory (Fabric Pipelines) and Dataflows Gen2—batch and (where relevant) near-real-time
- Develop robust SQL (T‑SQL/DAX‑aware patterns) and Spark notebooks (PySpark) for complex transformations
- Design semantic models that support BI and ML use cases; optimize for cost and performance (partitioning, indexing, caching, concurrency)
- Implement CDC/merge/upsert strategies for evolving datasets
- Implement data validation & testing (unit, schema, anomaly checks), monitor SLAs, and set up alerting & lineage
- Drive data reliability SLOs (freshness, completeness, accuracy)
- Enforce RBAC, row-level/column-level security, masking, and data retention
- Potentially integrate with Microsoft Purview for catalog, lineage, and policies
- Use Git integration in Fabric for version control and CI/CD (multi-environment deploys, templates, release automation)
- Establish IaC patterns where applicable (e.g., ARM/Bicep/Terraform for related Azure resources)
- Partner with Analytics Engineers, Data Scientists, and Power BI Developers to translate requirements into technical designs
- Mentor engineers and champion best practices (coding standards, reviews, documentation)
- Experience with Delta Lake patterns, medallion architecture, and near-real-time ingestion
- Familiarity with Power BI semantic models and performance implications
- Azure ecosystem exposure (Key Vault, Event Hub/Kafka, Azure Functions, Synapse migration experience)
- Certifications: Microsoft DP‑600 (Fabric Analytics Engineer Associate), DP‑203 (Azure Data Engineer Associate)
- Experience with regulated data (HIPAA, SOC 2, GDPR) and data contracts with product teams