Edit

Share via


Configure and manage data engineering and data science settings for Fabric capacities

Applies to: ✅ Data Engineering and Data Science in Microsoft Fabric

When you create Microsoft Fabric from the Azure portal, it's automatically added to the Fabric tenant that's associated with the subscription used to create the capacity. With the simplified setup in Microsoft Fabric, there's no need to link the capacity to the Fabric tenant. Because the newly created capacity will be listed in the admin settings pane. This configuration provides a faster experience for admins to start setting up the capacity for their enterprise analytics teams.

To make changes to the Data Engineering/Science settings in a capacity, you must have admin role for that capacity. To learn more about the roles that you can assign to users in a capacity, see Roles in capacities.

Use the following steps to manage the Data Engineering/Science settings for Microsoft Fabric capacity:

  1. Select the Settings option to open the setting pane for your Fabric account. Select Admin portal under Governance and insights section.

    Screenshot showing where to select Admin Portal settings.

  2. Choose the Capacity settings option to expand the menu and select the Fabric capacity tab. Here you should see the capacities that you have created in your tenant. Choose the capacity that you want to configure.

    Screenshot showing where to select Capacity settings.

  3. You're navigated to the capacity detail pane, where you can view the usage and other admin controls for your capacity. Navigate to the Data Engineering/Science Settings section and select Open Spark Compute. Configure the following parameters:

Note

At least one workspace should be attached to the Fabric Capacity to explore the Data Engineering/Science Settings from the Fabric Capacity Admin Portal.


Admin control: Disable Starter Pool usage

Capacity admins can now choose to disable Starter Pool usage across workspaces attached to the capacity. When disabled, users and workspace admins will no longer see the Starter Pool as a compute option. Instead, they must use custom pools explicitly created and managed by the capacity admin.

This feature provides centralized governance for compute usage, ensuring tighter control over compute sizing, cost, and scheduling behavior.

Tip

This setting is especially useful in large organizations that want to standardize compute patterns and avoid arbitrary consumption through default starter pools.


Admin control: Job-level bursting switch

Microsoft Fabric supports 3× bursting for Spark VCores, allowing a single job to temporarily use more compute cores than the base capacity provides. This improves job performance during bursts of activity by allowing full capacity utilization.

As a capacity admin, you can now control this behavior by using the "Disable job-level bursting" switch available in the Admin Portal:

  • Location:
    Admin Portal → Capacity Settings → [Select Capacity] → Data Engineering/Science Settings → Spark Compute

  • Behavior:

    • Enabled (Default): A single Spark job can consume the full burst limit (up to 3× Spark VCores).
    • Disabled: Individual Spark jobs are capped at the base capacity allocation, preserving concurrency and avoiding monopolization.

Note

This switch is only available when running Spark jobs on Fabric Capacity. If the Autoscale Billing option is enabled, this switch is automatically disabled because:

  • Autoscale Billing follows a pure pay-as-you-go model.
  • There's no smoothing window to allow usage bursts and balance them over 24 hours.
  • Bursting is a feature of reserved capacity, not on-demand autoscale compute.

Use cases and examples

Scenario Setting Behavior
Heavy ETL workload Bursting enabled (default) Job can use entire burst capacity (e.g., 384 Spark VCores in F64).
Multi-user interactive notebooks Bursting disabled Job usage is capped (e.g., 128 Spark VCores in F64), improving concurrency.
Autoscale Billing is enabled Bursting control unavailable All Spark usage is billed on demand; no bursting from base capacity.

Tip

Use this switch to optimize for either throughput or concurrency:

  • Keep bursting enabled for large jobs and pipelines.
  • Disable it for interactive or shared environments with many users.

Capacity Pools for Data Engineering and Data Science in Microsoft Fabric

  1. In the Pool List section of Spark Settings, click Add to create a Custom Pool for your Fabric Capacity.

    Screenshot showing the pool list section in Admin Portal settings.

    You're navigated to the pool creation page, where you can:

    • Specify the Pool name
    • Select Node family and Node size
    • Set Min and Max nodes
    • Enable/disable autoscale and dynamic allocation of executors

    Screenshot showing the pool creation section in Admin Portal settings.

  2. Select Create to save your settings.

    Screenshot showing the capacity pools saved in Admin Portal settings.

Note

Capacity-level custom pools have a startup latency of 2–3 minutes. For faster Spark session startup (<5 seconds), use Starter Pools if enabled.

  1. Once created, the capacity pool becomes available in:

    • The Pool Selection dropdown in workspace settings
    • The Environment Compute Settings page in workspaces

    Screenshot showing the capacity pools listed in Workspace settings.

    Screenshot showing the capacity pools listed in Environment settings.

  2. This enables centralized compute governance. Admins can create standardized pools, and optionally disable workspace-level customization, preventing admins in workspaces from modifying pool settings or creating their own.