Job Title : Fabric Data Engineer
This role offers an exciting opportunity to work on a high-profile project for a US health insurer, where you will be responsible for standing up and managing the Fabric Mirroring from SQL Server and other sources into Onelake. You will own the process of CDC (Change Data Capture) and schema drift at scale, designing resilient high-volume ingestion pipelines that teams can trust.
Main Responsibilities :
- Stand up and manage Fabric Mirroring from SQL Server and other sources into Onelake.
- Handle change at scale, including soft hard deletes, late arriving data, and backfills using reliable watermarking and reconciliation checks.
- Design resilient ingestion pipelines with Fabric Data Factory and / or notebooks, adding retries, dead-lettering, and circuit-breaker patterns for fault tolerance.
- Manage schema drift, automating drift detection and schema evolution, publishing change notes and guardrails so downstream consumers aren't surprised.
- Performance cost tuning, optimising batch sizes, file counts, partitions, parallelism, and capacity usage to balance speed, reliability, and spend.
- Observability quality, instrumenting lineage, logs, metrics, and DQ tests, nulls ranges uniqueness, setting up alerts and simple SLOs for ingestion health.
Must-Have Skills :
SQL Server T-SQL; CDC / replication fundamentals.Microsoft Fabric Mirroring; Onelake / Lakehouse; Onelake shortcuts.Schematic shift detection / management and data contracts.Familiarity with large complex relational databases.Python Scala / Spark for ingestion and validation.Git-based workflow basic CI / CD ( Fabric Deployment Pipelines or Azure DevOps).