The Problem Microsoft Fabric Solves
Enterprise analytics has suffered from fragmentation for over a decade. A typical enterprise analytics stack includes a data lake for raw data storage, an ETL tool for data movement and transformation, a data warehouse for structured analytics, a real-time streaming platform for event processing, a data science environment for machine learning, and a business intelligence tool for reporting and dashboards. Each component has its own provisioning, security model, governance framework, billing, and management overhead.
The result is what the industry calls the "analytics tax" — the operational cost of managing integration between these components, maintaining consistent security and governance, and troubleshooting data pipeline failures across service boundaries. For enterprise organizations, this tax often consumes 40-60% of the analytics team's time, leaving the minority for actual analysis and insight generation.
Microsoft Fabric addresses this problem by unifying all analytics workloads on a single platform with a shared data layer (OneLake), shared security model, shared governance, and shared compute capacity. Data is stored once and accessed by all workloads without movement or duplication. This architectural simplification is not incremental — it fundamentally changes how enterprise analytics teams operate.
Understanding OneLake: The Foundation of Fabric
OneLake is Fabric's built-in data lake, and understanding it is essential to understanding Fabric's value proposition. OneLake is to Fabric what OneDrive is to Microsoft 365 — a single, unified storage layer that every workload uses by default.
Key architectural characteristics of OneLake include automatic provisioning with every Fabric tenant (no separate storage account setup), hierarchical organization using workspaces, lakehouses, and folders, native Delta Lake (Parquet) format for all structured and semi-structured data, one copy of data serving all analytics workloads (no ETL between components), ADLS Gen2 API compatibility for existing tools and scripts, and built-in governance through Microsoft Purview integration.
OneLake also supports shortcuts — virtual references to data in external storage accounts (Azure Data Lake Storage, Amazon S3, Google Cloud Storage) that make external data appear as native OneLake data without copying. This is particularly valuable during migration, allowing organizations to gradually move data to OneLake while maintaining access to existing storage.
For enterprise organizations, OneLake eliminates the most expensive aspect of traditional analytics architectures: data movement. When every workload reads from the same data store, there are no ETL pipelines to build and maintain between the data lake and data warehouse, no data synchronization issues between the BI tool and the analytics database, no security model inconsistencies between different storage systems, and no duplicate storage costs for the same data in multiple formats.
Fabric Workloads: A Deep Dive
Data Engineering
The Data Engineering workload provides Apache Spark-based processing for large-scale data transformation, cleansing, and preparation. This is where raw data from source systems is ingested, cleaned, enriched, and shaped into analytics-ready formats.
Fabric's Spark implementation runs on a managed Spark runtime — no cluster provisioning, no Spark version management, no infrastructure configuration. Engineers write PySpark, Spark SQL, Scala, or R code in notebooks, and Fabric handles the compute allocation from the capacity pool. Starter pools provide near-instant Spark session startup (under 30 seconds), eliminating the 3-5 minute cold start that plagued Azure Synapse Spark pools.
Data Factory pipelines are integrated directly into Fabric for orchestration. This includes 150+ connectors for data ingestion, support for complex pipeline patterns (conditional logic, loops, error handling), monitoring and alerting for pipeline failures, and incremental refresh patterns for large datasets. For organizations currently using Azure Data Factory, the migration to Fabric pipelines is straightforward — the pipeline authoring experience is nearly identical, and many existing pipelines can be migrated with minimal modification.
Data Warehouse
The Fabric Warehouse provides a fully managed, T-SQL-compatible analytical database that reads and writes data directly to OneLake in Delta format. This is the entry point for organizations with strong SQL skills who want the benefits of Fabric without immediately adopting Spark.
Warehouse capabilities include full T-SQL DML (INSERT, UPDATE, DELETE, MERGE), stored procedures and functions, cross-warehouse and cross-database queries, automatic statistics and query optimization, workload isolation for concurrent queries, and visual query designer for no-code query building. The critical distinction between Fabric Warehouse and the Lakehouse SQL endpoint is write capability. The Lakehouse SQL endpoint provides read-only T-SQL access to Delta tables. The Warehouse supports full read-write operations, making it suitable for traditional data warehouse patterns including staging, transformation, and dimension management.
Lakehouse
The Lakehouse combines the flexibility of a data lake with the structure of a data warehouse. Data is stored in Delta Lake format, providing ACID transactions, schema enforcement, and time travel on top of Parquet files. The Lakehouse is the recommended starting point for organizations building new analytics platforms in Fabric.
Each Lakehouse provides a managed file area for unstructured and semi-structured data, a managed tables area with Delta tables, an automatic SQL endpoint for T-SQL queries, and a default semantic model for Power BI. This dual-mode access means data engineers work with Spark and analysts work with SQL — against the same data, with the same governance, and with no data movement between them.
Data Science
The Data Science workload provides a managed environment for machine learning that integrates MLflow for experiment tracking, model registration, and deployment. Data scientists work in the same Spark notebooks as data engineers, with access to the same data in OneLake.
Key capabilities include native MLflow integration for experiment tracking and model versioning, pre-installed ML libraries (scikit-learn, TensorFlow, PyTorch, XGBoost), PREDICT function for applying trained models in T-SQL queries, and SynapseML for distributed machine learning at scale. For enterprise organizations, the governance integration is the key differentiator. Models are registered, versioned, and governed through the same framework as data assets, providing a complete lineage from raw data through training data to model to prediction.
Real-Time Analytics
Real-Time Analytics (based on Azure Data Explorer / Kusto) provides sub-second query performance on streaming and time-series data. This workload handles IoT telemetry, application logs, clickstream data, and any scenario requiring near-instant analysis of high-volume event streams.
The Eventstream component provides a no-code interface for configuring real-time data ingestion from Azure Event Hubs, Kafka, custom applications, and other streaming sources. Data flows through Eventstreams into KQL databases where it is immediately queryable using Kusto Query Language (KQL). Real-Time Dashboards provide live visualizations that update automatically as new data arrives.
Power BI Integration
Power BI in Fabric is not a bolted-on addition — it is a native workload that reads directly from OneLake. The most significant advancement is Direct Lake mode, which provides import-mode performance with DirectQuery-level data freshness. Direct Lake loads Delta table column chunks directly into memory, bypassing the traditional import process that required scheduled refreshes.
For enterprise organizations, this means dashboards that always show current data without the refresh lag that plagued import-mode datasets. Reports on multi-billion-row datasets respond in seconds. And because there is no import process, there is no duplication of data between OneLake and the Power BI model — reducing storage costs and eliminating refresh failures.
Fabric Licensing and Capacity Planning
| SKU | Capacity Units | Approx. Monthly Cost | Recommended For |
|---|---|---|---|
| F2 | 2 CU | ~$262 | Development and testing |
| F64 | 64 CU | ~$5,242 | Small-to-mid enterprise production |
| F128 | 128 CU | ~$10,483 | Mid enterprise with multiple workloads |
| F256 | 256 CU | ~$20,966 | Large enterprise with heavy workloads |
| F512 | 512 CU | ~$41,932 | Enterprise with large-scale analytics |
Critical licensing considerations for enterprise organizations. F64 is the minimum SKU that includes Power BI Pro equivalent capabilities for all users in the capacity. Below F64, users still need individual Power BI Pro licenses ($10/user/month). Capacity is shared across all workloads — a heavy Spark job reduces available capacity for concurrent Power BI queries. Capacity auto-scaling is available to handle burst workloads, and reservations (1-year or 3-year) provide significant discounts (up to 40%) over pay-as-you-go pricing.
For cost comparison, consider a typical enterprise running Azure Data Factory ($500/month), Synapse dedicated SQL pool ($3,000/month), Synapse Spark pool ($2,000/month), and Power BI Premium P1 ($4,995/month). That is approximately $10,495/month for separate services. An F128 Fabric capacity at $10,483/month provides equivalent or greater capability for all four workloads in a single, unified platform.
Enterprise Governance in Microsoft Fabric
Governance is where Fabric delivers perhaps its most significant enterprise value. Traditional analytics environments suffer from governance fragmentation — different security models for the data lake versus the warehouse, separate audit logs for different services, and no unified lineage tracking across the analytics pipeline.
Fabric governance capabilities include workspace-level access control with role-based permissions, item-level security for individual data assets, row-level and column-level security in Warehouses and Lakehouses, sensitivity labels from Microsoft Purview applied across all workloads, unified audit logging for all data access and modifications, end-to-end data lineage from source through transformation to report, endorsement labels (Certified, Promoted) for trusted data assets, and domain-based organization for multi-department environments.
For regulated industries, Fabric's governance model provides the unified audit trail and access control framework that compliance officers require. A single audit log covers all data access across all workloads, eliminating the log correlation challenges that plague multi-service analytics architectures.
Migrating from Azure Synapse to Microsoft Fabric
Microsoft has signaled clearly that Fabric is the future of its analytics platform strategy. Azure Synapse Analytics continues to be supported, but new feature investment is focused on Fabric. Enterprise organizations should begin planning their migration strategy now.
Migration Path for Synapse Dedicated SQL Pools
Synapse dedicated SQL pools map most directly to Fabric Warehouses. The T-SQL surface area is highly compatible, though some features differ. Migration involves exporting data from dedicated SQL pools to Parquet format in ADLS Gen2, creating Fabric Warehouses and establishing OneLake shortcuts to the exported data, recreating stored procedures, views, and security configurations, validating query performance and results parity, and redirecting BI tools and applications to the Fabric Warehouse endpoints.
Migration Path for Synapse Spark Pools
Synapse Spark notebooks can generally be used in Fabric with minimal modification. The primary changes involve updating storage paths from ADLS Gen2 to OneLake references, migrating linked services to Fabric connections, adjusting Spark configurations for Fabric's managed runtime, and reconfiguring any Azure Key Vault integrations. The biggest benefit of migration is elimination of Spark pool management — no more worrying about pool sizing, auto-scale configuration, or library management.
Migration Path for Azure Data Factory
Data Factory pipelines in Fabric use the same authoring experience as standalone Azure Data Factory. Migration involves recreating pipelines in Fabric (or using export/import for simple pipelines), migrating linked services to Fabric connections, updating dataset references to OneLake, and reconfiguring triggers and monitoring. Organizations with complex ADF estates (hundreds of pipelines) should plan for a phased migration over 3-6 months.
How EPC Group Helps with Microsoft Fabric
As a Microsoft Gold Partner with deep expertise in Power BI and enterprise analytics, EPC Group provides end-to-end Fabric consulting services. Our team has been working with Fabric since its public preview and has completed Fabric implementations for Fortune 500 organizations across healthcare, finance, and government.
- Fabric readiness assessments — Evaluate your current analytics estate, identify migration candidates, and develop a prioritized adoption roadmap
- Architecture design — Design workspace structures, security models, and data architectures optimized for your workloads and compliance requirements
- Migration execution — Hands-on migration from Azure Synapse, Databricks, or other platforms to Fabric, including data validation and performance testing
- Capacity planning and optimization — Right-size your Fabric capacity, implement monitoring, and optimize workload scheduling to maximize cost efficiency
- Training and enablement — Role-based training for data engineers, analysts, and administrators to ensure your team can operate Fabric independently
Frequently Asked Questions
What is Microsoft Fabric and how is it different from Azure Synapse?
Microsoft Fabric is a unified SaaS analytics platform that brings together data engineering, data science, real-time analytics, data warehousing, and business intelligence in a single product with a shared data lake (OneLake). Unlike Azure Synapse, which required separate provisioning and management of individual services (Synapse SQL, Spark pools, Data Explorer), Fabric provides all workloads in a single capacity with unified governance, security, and billing. Fabric eliminates the need to manage infrastructure, move data between services, or maintain separate security models for each analytics component. Azure Synapse Analytics is being incorporated into Fabric, with Microsoft encouraging migration to the Fabric platform.
How much does Microsoft Fabric cost for enterprise organizations?
Microsoft Fabric uses capacity-based licensing measured in Capacity Units (CUs). The F2 SKU starts at approximately $262/month, while enterprise organizations typically require F64 ($5,242/month) or higher. F256 capacity runs approximately $20,966/month and F512 at approximately $41,932/month. Power BI Pro is included in F64 and above capacities. Organizations can also use pay-as-you-go pricing for burst workloads. The key cost advantage is consolidation — instead of paying separately for Azure Data Factory, Synapse, Data Explorer, and Power BI Premium, all workloads run on a single capacity with shared compute resources. Trial capacities are available for evaluation.
Can Microsoft Fabric replace our existing data warehouse?
Yes, Microsoft Fabric can replace traditional data warehouses for most enterprise use cases. Fabric offers two primary storage paradigms: the Lakehouse (based on Delta Lake format, optimized for large-scale data engineering and data science) and the Warehouse (T-SQL-based, optimized for traditional BI workloads and SQL-familiar teams). The Warehouse supports full T-SQL DML operations, stored procedures, and cross-database queries. For organizations currently on Azure SQL Data Warehouse or Synapse dedicated SQL pools, migration to Fabric Warehouse provides comparable functionality with simplified management, unified governance, and integrated Power BI. However, organizations with extremely specialized analytical database requirements (columnar compression, materialized views at massive scale) should conduct a thorough proof-of-concept before committing to migration.
How does Microsoft Fabric integrate with Power BI?
Power BI is natively integrated into Microsoft Fabric as one of its core workloads. All data stored in OneLake is automatically accessible to Power BI without data movement or duplication. Direct Lake mode allows Power BI to query Delta tables in OneLake directly, combining the performance of import mode with the freshness of DirectQuery. Semantic models (formerly datasets) can be built directly on Fabric lakehouses and warehouses. Capacity purchased for Fabric also covers Power BI workloads — organizations on F64 or above get Power BI Pro equivalent capabilities included. This integration eliminates the historical gap between data engineering and business intelligence, enabling analysts to build reports directly on governed, engineering-quality data assets.
What skills does our team need to use Microsoft Fabric effectively?
Microsoft Fabric is designed to serve multiple skill levels. Data engineers work primarily with Spark notebooks (PySpark, Spark SQL, Scala) and Data Factory pipelines for ETL orchestration. Data scientists use Spark notebooks with ML libraries and the built-in MLflow integration for experiment tracking. SQL analysts work with the Warehouse experience using familiar T-SQL syntax. Business analysts create reports using Power BI, which can now connect directly to Fabric data assets. Data administrators manage governance through Purview integration, security policies, and monitoring dashboards. The most critical skill gap for most organizations is Spark/PySpark proficiency for the Lakehouse workload — this is where training investment delivers the highest return. Organizations heavily invested in T-SQL skills can start with the Warehouse workload while gradually building Spark capabilities.
Ready to Explore Microsoft Fabric?
EPC Group provides Fabric readiness assessments, architecture design, and migration services for enterprise organizations. Start with a complimentary consultation to evaluate whether Fabric is the right platform for your analytics workloads.
Schedule a Fabric ConsultationErrin O'Connor
CEO & Chief AI Architect at EPC Group | 28+ years Microsoft consulting | Author, Power BI Field Guide (Microsoft Press)