EPC Group - Enterprise Microsoft AI, SharePoint, Power BI, and Azure Consulting
G2 High Performer Summer 2025, Momentum Leader Spring 2025, Leader Winter 2025, Leader Spring 2026
BlogContact
Ready to transform your Microsoft environment?Get started today
(888) 381-9725Get Free Consultation
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌
‌

EPC Group

Enterprise Microsoft consulting with 28+ years serving Fortune 500 companies.

(888) 381-9725
contact@epcgroup.net
4900 Woodway Drive - Suite 830
Houston, TX 77056

Follow Us

Solutions

  • All Services
  • Microsoft 365 Consulting
  • AI Governance
  • Azure AI Consulting
  • Cloud Migration
  • Microsoft Copilot
  • Data Governance
  • Microsoft Fabric
  • vCIO / vCAIO Services
  • Large-Scale Migrations
  • SharePoint Development

Industries

  • All Industries
  • Healthcare IT
  • Financial Services
  • Government
  • Education
  • Teams vs Slack

Power BI

  • Case Studies
  • 24/7 Emergency Support
  • Dashboard Guide
  • Gateway Setup
  • Premium Features
  • Lookup Functions
  • Power Pivot vs BI
  • Treemaps Guide
  • Dataverse
  • Power BI Consulting

Company

  • About Us
  • Our History
  • Microsoft Gold Partner
  • Case Studies
  • Testimonials
  • Blog
  • Resources
  • Contact

Microsoft Teams

  • Teams Questions
  • Teams Healthcare
  • Task Management
  • PSTN Calling
  • Enable Dial Pad

Azure & SharePoint

  • Azure Databricks
  • Azure DevOps
  • Azure Synapse
  • SharePoint MySites
  • SharePoint ECM
  • SharePoint vs M-Files

Comparisons

  • M365 vs Google
  • Databricks vs Dataproc
  • Dynamics vs SAP
  • Intune vs SCCM
  • Power BI vs MicroStrategy

Legal

  • Sitemap
  • Privacy Policy
  • Terms
  • Cookies

Our Specialized Practices

PowerBIConsulting.com|CopilotConsulting.com|SharePointSupport.com

© 2026 EPC Group. All rights reserved.

Microsoft Fabric Data Engineering - EPC Group enterprise consulting

Microsoft Fabric Data Engineering

Enterprise Guide 2026: Lakehouse architecture, Spark notebooks, data pipelines, Delta Lake, and governance best practices for production-grade data engineering.

What Is Data Engineering in Microsoft Fabric?

Featured Snippet: Data engineering in Microsoft Fabric is a unified discipline that combines lakehouse architecture, Apache Spark notebooks, Data Factory pipelines, and Delta Lake storage into a single SaaS platform. Fabric eliminates the need to stitch together separate Azure services — data engineers get OneLake storage, Spark compute, orchestration, and governance in one experience with zero infrastructure management.

Microsoft Fabric has fundamentally changed how enterprise data engineering teams build, orchestrate, and govern data platforms. Before Fabric, building an enterprise data pipeline on Azure required provisioning and integrating at least four separate services: Azure Data Lake Storage Gen2 for storage, Azure Synapse or HDInsight for Spark processing, Azure Data Factory for orchestration, and Power BI for downstream analytics. Each service had its own security model, billing structure, and operational overhead.

Fabric collapses this entire stack into a single, capacity-based SaaS experience. Data engineers work within a Lakehouse that combines the flexibility of a data lake with the structure of a data warehouse. They write transformations in Spark notebooks, orchestrate workflows with Data Factory pipelines, and the results flow directly into Power BI — all on the same OneLake storage layer. No data movement between services, no separate access control configurations, no cluster management.

This guide covers every aspect of Fabric data engineering that enterprise teams need to master in 2026: Lakehouse architecture, Spark notebooks, data pipelines, shortcuts, Delta Lake optimization, the medallion pattern, performance tuning, governance, and cost management. Whether you are migrating from Azure Synapse, Databricks, or building greenfield — this is your comprehensive reference.

Fabric Lakehouse Architecture

The Fabric Lakehouse is the foundational construct for data engineering. Unlike a traditional data warehouse that requires upfront schema design, or a data lake that stores files without structure, the Lakehouse provides both: schema-on-write for structured tables and schema-on-read for raw file ingestion, unified on a single storage layer.

Every Lakehouse in Fabric automatically provisions two endpoints: a Spark endpoint for notebook-based processing and a SQL analytics endpoint for T-SQL queries. Both endpoints operate on the same Delta tables in OneLake — there is no data duplication or ETL required between them. This dual-endpoint architecture means data engineers can write Spark transformations while analysts simultaneously query the same tables with SQL, each using their preferred tool.

OneLake Storage

Single data lake for the entire organization. All Fabric workspaces share OneLake, eliminating data silos and redundant copies across teams.

Delta Lake Format

All Lakehouse tables use Delta format — ACID transactions, time travel, schema evolution, and Z-order optimization built into every table.

Dual Endpoints

Every Lakehouse exposes Spark and SQL analytics endpoints simultaneously. No data movement — one table, two access patterns.

Files + Tables

Store unstructured files (CSV, JSON, images) in the Files section and structured Delta tables in the Tables section — same Lakehouse.

For enterprise deployments, EPC Group recommends a multi-Lakehouse pattern: one Lakehouse per domain (sales, finance, operations) within a shared workspace, with cross-Lakehouse references via shortcuts. This provides domain isolation for security and governance while maintaining a unified data fabric. Each domain team manages their own Lakehouse lifecycle — schema changes, access policies, and data quality — without impacting other teams.

Notebooks and Apache Spark in Fabric

Fabric Notebooks are the primary tool for data transformation. They run Apache Spark on Microsoft-managed compute pools — no cluster provisioning, no Spark configuration, no node management. When you open a notebook and execute a cell, Fabric automatically allocates Spark resources from your capacity and releases them when the session ends. This serverless model eliminates the operational overhead that makes Spark notoriously difficult to manage in traditional environments.

Notebooks support PySpark, Spark SQL, Scala, and R. For most enterprise data engineering, PySpark and Spark SQL cover 95% of use cases. Fabric also supports notebook parameterization — you can call notebooks from pipelines with runtime parameters, enabling the same transformation logic to process different datasets, date ranges, or environments without code duplication.

Notebook Best Practices for Enterprise

  • Use notebook resources to store shared utility functions and import them across notebooks — eliminates code duplication.
  • Set session timeouts (default 20 minutes idle) to prevent capacity waste from forgotten sessions.
  • Pin Spark libraries at the workspace level, not per-notebook, for consistent dependency management across teams.
  • Use mssparkutils for credential management — never hardcode storage keys or connection strings in notebook cells.
  • Structure notebooks with clear markdown sections: parameters, imports, transformations, validation, output — for readability and debugging.
  • Enable V-Order on write operations (spark.sql.parquet.vorder.enabled = true) to optimize downstream DirectLake performance.
  • Use the high-concurrency mode for development environments — shares a Spark session across multiple notebooks, reducing startup time.

Data Pipelines and Orchestration

Fabric Data Factory pipelines are the orchestration backbone for enterprise data engineering. They provide a visual, drag-and-drop interface for building multi-step data workflows that ingest from source systems, execute Spark notebook transformations, run stored procedures, and trigger downstream refreshes. For teams familiar with Azure Data Factory, Fabric pipelines are architecturally identical — same activity types, same expression language, same linked service model.

The key difference from standalone ADF is integration depth. Fabric pipelines natively reference Lakehouse tables, notebooks, and semantic models within the same workspace — no connection strings or linked services required. A pipeline can copy data from an external SQL Server into a Lakehouse table, trigger a Spark notebook to transform it, and refresh a Power BI semantic model, all in a single orchestrated workflow with full lineage tracking.

90+ Connectors

Ingest from SQL Server, Oracle, Salesforce, SAP, REST APIs, file systems, and cloud storage — same connector library as Azure Data Factory.

Dependency Management

Chain activities with success/failure/completion dependencies. Build complex DAGs with parallel branches and conditional logic.

Scheduling & Triggers

Schedule pipelines on cron expressions, tumbling windows, or event triggers. Support for parameterized schedules across environments.

For enterprise-scale orchestration, EPC Group recommends a hub-and-spoke pipeline pattern: a master pipeline that calls child pipelines per data domain (sales, finance, HR). Each child pipeline encapsulates the full bronze-silver-gold transformation for its domain. The master pipeline manages cross-domain dependencies and sends consolidated alerting on success or failure. This pattern scales cleanly as new data domains are onboarded and simplifies debugging by isolating failures to specific domains.

OneLake Shortcuts: Federated Data Access

Shortcuts are one of Fabric's most powerful and underutilized data engineering features. A shortcut is a reference pointer to data that lives outside your Lakehouse — in another Lakehouse, in ADLS Gen2, in Amazon S3, or in Google Cloud Storage. The data appears as a native table or folder in your Lakehouse, queryable via Spark and SQL, but it is never physically copied to OneLake. You pay zero OneLake storage for shortcut data.

For enterprise data engineering, shortcuts solve three critical challenges: data residency (data stays in its regulated location), cost optimization (no storage duplication), and incremental migration (reference legacy storage while building new pipelines). Shortcuts also enable cross-workspace and cross-tenant data sharing — a finance team can create a shortcut to the sales team's gold-layer tables without requesting a data copy.

Shortcut Use Cases in Enterprise

  • Multi-cloud federation: Reference S3 or GCS data from Fabric without migrating storage to Azure.
  • Compliance boundaries: Keep HIPAA-regulated data in its approved ADLS Gen2 account while analyzing it in Fabric.
  • Cross-domain data mesh: Each domain publishes gold-layer tables, other domains consume via shortcuts.
  • Migration bridge: Shortcut legacy ADLS data into Fabric Lakehouse while incrementally rebuilding pipelines.
  • Cost avoidance: Reference a 10 TB dataset via shortcut instead of duplicating it — saves ~$230/month in storage alone.

Delta Lake and Medallion Architecture

Every table in a Fabric Lakehouse is a Delta Lake table. Delta Lake adds ACID transactions, schema enforcement, time travel (versioned history), and merge operations (upserts) on top of Parquet files. For data engineers, this means you get warehouse-grade reliability on lake-scale storage — no more corrupted partial writes, no schema drift breaking downstream reports, no inability to roll back a bad transformation.

The medallion architecture is the recommended pattern for organizing Delta tables within a Fabric Lakehouse. It provides a clear, auditable data lineage from raw ingestion to business-ready analytics. Each layer has a specific purpose, quality standard, and access pattern.

Bronze (Raw)

Ingest raw data exactly as received from source systems. Append-only, no transformations. Add ingestion metadata columns (source_system, ingestion_timestamp, batch_id). This layer serves as the immutable audit trail — you can always reprocess from bronze if silver/gold logic changes.

Silver (Conformed)

Apply cleansing, deduplication, type casting, null handling, and business key resolution. Enforce schema with Delta schema enforcement. Merge (upsert) patterns for slowly changing dimensions. Silver tables are the single source of truth for conformed enterprise data.

Gold (Business)

Produce aggregated, denormalized, business-ready tables optimized for Power BI DirectLake mode. Star schema design with fact and dimension tables. V-Order optimized for maximum query performance. Gold tables serve analysts, reports, and AI/ML feature engineering.

Performance Optimization

Fabric data engineering performance depends on three factors: how you write data (table optimization), how you process data (Spark tuning), and how downstream consumers read data (DirectLake compatibility). Getting all three right is the difference between a platform that delivers sub-second dashboards and one that frustrates analysts with spinning wheels.

Performance Optimization Checklist

  • Enable V-Order on all gold-layer tables — this Fabric-specific optimization sorts data for maximum DirectLake query performance.
  • Run OPTIMIZE regularly on Delta tables to compact small files — target 128 MB per file for optimal Spark read performance.
  • Apply Z-ORDER on columns frequently used in WHERE clauses — enables data skipping and reduces scan volume by 80%+.
  • Use table partitioning by date for large fact tables (>100 GB) — but avoid over-partitioning on high-cardinality columns.
  • Set spark.sql.shuffle.partitions based on data volume — default 200 is too high for small tables, too low for billion-row jobs.
  • Use Delta table VACUUM to remove old file versions — reduces storage cost and improves file listing performance.
  • Broadcast small dimension tables in joins (< 100 MB) to avoid expensive shuffle operations.
  • Cache intermediate DataFrames that are used multiple times in a notebook — avoids recomputation.

Data Engineering Governance

Governance in Fabric data engineering is not an afterthought — it is built into the platform at every layer. OneLake provides automatic lineage tracking from source ingestion through bronze-silver-gold transformations to Power BI reports. Microsoft Purview integrates natively with Fabric to classify sensitive data, apply sensitivity labels, and enforce access policies across the entire data estate.

For enterprise data engineering teams, governance manifests in four key areas: access control (who can read/write which Lakehouse tables), data classification (what sensitivity level does each column contain), lineage (where did this data come from and what transformations were applied), and quality (are the data values accurate, complete, and timely). Fabric addresses the first three natively; data quality requires engineering discipline through validation notebooks and monitoring.

Access Control

Workspace roles (Admin, Member, Contributor, Viewer) control Lakehouse access. Row-level security and object-level security for fine-grained table protection.

Data Classification

Purview automatically scans Lakehouse tables for PII, financial data, and health information. Sensitivity labels propagate from source to downstream artifacts.

Lineage Tracking

Automatic lineage from pipeline ingestion through notebook transformations to Power BI reports. No manual documentation required — Fabric tracks every dependency.

Quality Monitoring

Build validation notebooks that check row counts, null rates, schema conformance, and business rule compliance after each pipeline run.

Cost Management for Fabric Data Engineering

Fabric uses a capacity-based pricing model — you purchase Capacity Units (CUs) that are shared across all Fabric workloads in a capacity. Data engineering consumes CUs when Spark notebooks execute, when pipelines run copy activities, and when Dataflow Gen2 transformations process data. OneLake storage is billed separately at approximately $0.023 per GB per month (same as ADLS Gen2 hot tier).

The most common cost mistake in Fabric data engineering is over-provisioning capacity for development environments. A single F64 capacity ($4,096/month reserved) provides more than enough compute for a 10-person data engineering team running development and testing workloads. Production workloads with heavy Spark processing may need F128 or F256, but EPC Group recommends starting with F64 and scaling based on actual utilization metrics rather than estimated workload projections.

Cost Optimization Quick Wins

  • Reserved capacity saves 40% vs PAYG — commit to F64+ for 1 year on any production workload.
  • Spark session timeouts: Set to 5-10 minutes for dev, 2 minutes for production to reclaim idle capacity.
  • Schedule heavy batch jobs during off-peak hours (nights/weekends) when interactive usage is low.
  • Use shortcuts for reference data — a 5 TB shortcut costs $0 in OneLake storage vs $115/month for a copy.
  • VACUUM Delta tables weekly — removes old versions, reducing storage by 20-50% on active tables.
  • Monitor capacity utilization in the Fabric admin portal — sustained >80% means you need to scale up or optimize.

Frequently Asked Questions: Fabric Data Engineering

What is data engineering in Microsoft Fabric?

Data engineering in Microsoft Fabric is a unified discipline that combines lakehouse architecture, Apache Spark notebooks, data pipelines, and Delta Lake storage into a single SaaS platform. Fabric data engineers use OneLake as the centralized storage layer, write transformations in PySpark or Spark SQL notebooks, orchestrate workflows with Data Factory pipelines, and leverage shortcuts to connect external data sources — all without managing infrastructure. The result is a modern data engineering experience that eliminates the complexity of stitching together separate Azure services like Synapse, Data Factory, and ADLS Gen2.

What is the Fabric Lakehouse and how does it work?

The Fabric Lakehouse is a combined data lake and data warehouse that stores data in open Delta Lake (Parquet) format on OneLake. It supports both SQL analytics and Spark-based data engineering on the same data without duplication. You create tables that are automatically registered in the SQL analytics endpoint for T-SQL queries and simultaneously accessible via Spark notebooks for transformations. The Lakehouse eliminates the traditional choice between a data lake (flexible but ungoverned) and a data warehouse (structured but rigid) — delivering both capabilities on a single copy of data.

How do Fabric Notebooks compare to Databricks Notebooks?

Fabric Notebooks run Apache Spark on Microsoft-managed compute — no cluster provisioning or configuration required. They support PySpark, Spark SQL, Scala, and R with built-in visualization and collaboration features. Databricks Notebooks offer more advanced features like MLflow integration, Databricks Connect for local IDE development, and more granular cluster control. For standard data engineering workloads (ETL, data cleansing, aggregation), Fabric Notebooks are equally capable with significantly lower operational overhead. For advanced ML engineering and custom Spark tuning, Databricks retains an edge.

What are Fabric Shortcuts and when should I use them?

Fabric Shortcuts are pointers to external data sources that make data appear as if it lives in your Lakehouse without physically copying it. Shortcuts support ADLS Gen2, Amazon S3, Google Cloud Storage, and Dataverse. Use shortcuts when: (1) data must remain in its source system for compliance, (2) you want to avoid storage duplication costs, (3) you need to federate data across organizational boundaries, or (4) you are migrating incrementally and want to reference legacy storage during transition. Shortcuts are read-only by default and respect the security policies of the source system.

What is the medallion architecture in Microsoft Fabric?

The medallion architecture (bronze-silver-gold) is the recommended data organization pattern in Fabric Lakehouse. Bronze layer ingests raw data from source systems with minimal transformation — preserving the original format for auditability. Silver layer applies cleansing, deduplication, schema enforcement, and business logic to create conformed datasets. Gold layer produces aggregated, business-ready tables optimized for reporting and analytics. In Fabric, each layer is a set of Delta tables in the Lakehouse, with Spark notebooks or Data Factory pipelines orchestrating the transformations between layers.

How does Fabric handle data pipeline orchestration?

Fabric uses Data Factory pipelines for orchestration — a visual, code-free interface for scheduling and sequencing data movement and transformation activities. Pipelines support 90+ connectors for ingesting data from cloud and on-premises sources. You can chain Spark notebook executions, stored procedures, Dataflow Gen2 transformations, and copy activities into multi-step workflows with dependency management, retry logic, and alerting. Pipelines also support parameterization, allowing the same pipeline to process different datasets or environments dynamically.

What are the cost management strategies for Fabric data engineering?

Key cost strategies include: (1) Use reserved capacity (F64+) for predictable workloads — saves 40% vs pay-as-you-go. (2) Implement workspace-level capacity assignment to isolate cost by team or project. (3) Use Spark session timeouts to prevent idle compute consumption. (4) Leverage V-Order optimization on Delta tables to reduce query compute. (5) Schedule heavy pipelines during off-peak hours to smooth capacity utilization. (6) Use shortcuts instead of data copies to eliminate redundant storage. (7) Monitor capacity metrics in the Fabric admin portal and set alerts for sustained high utilization.

How does governance work for Fabric data engineering?

Fabric governance is built on Microsoft Purview integration and OneLake security. Data engineers benefit from: automatic data lineage tracking across notebooks and pipelines, sensitivity labels that propagate from source to downstream tables, workspace-level access control with Entra ID, row-level and object-level security on Lakehouse tables, and endorsement workflows (certified/promoted) for dataset quality signaling. All data in OneLake is encrypted at rest and in transit. Purview Data Catalog automatically discovers and classifies Lakehouse tables, enabling data stewards to manage the entire engineering lifecycle from a single governance plane.

Can I migrate my existing Azure Synapse or ADF workloads to Fabric?

Yes. Microsoft provides migration paths from Azure Synapse Analytics and Azure Data Factory to Fabric. Synapse Spark pools map directly to Fabric Spark notebooks with minimal code changes. ADF pipelines can be migrated to Fabric Data Factory with the pipeline migration wizard — most activities transfer directly. Synapse SQL dedicated pools require more effort, as Fabric uses a different SQL engine. EPC Group recommends a phased migration: start with new workloads on Fabric, migrate existing ADF pipelines next, then gradually transition Synapse workloads as Fabric capabilities mature. We have completed 50+ Fabric migrations for enterprise clients.

Related Resources

Microsoft Fabric Consulting Services

Enterprise Fabric implementation, migration, and optimization services from EPC Group.

Read more

Microsoft Fabric Enterprise Guide

Comprehensive overview of Microsoft Fabric capabilities, licensing, and adoption strategy.

Read more

Fabric vs Databricks Comparison

Head-to-head comparison of Microsoft Fabric and Databricks for enterprise data platforms.

Read more

Need Help Building Your Fabric Data Engineering Platform?

EPC Group has completed 50+ Microsoft Fabric implementations for enterprise clients. From Lakehouse architecture design to production pipeline deployment, our certified Fabric engineers deliver data platforms that scale. Schedule a free data engineering assessment today.

Get Fabric Assessment (888) 381-9725