Share via

Monitoring Microsoft Foundry Models Using Diagnostic Settings

Rushikesh Gaidhani 161 Reputation points
2026-04-02T18:51:38.33+00:00

Hello,

I created a Microsoft Foundry resource and deployed a few OpenAI models. Although I enabled Diagnostic Settings and routed AllMetrics to Log Analytics, the metrics currently appear only for the parent Foundry resource rather than for each model deployment.

I can see individual model metrics in Azure Monitor > Metrics when I split by Model Name.

What is the recommended way to capture and monitor per-model deployment metrics?
User's image

Azure OpenAI Service
Azure OpenAI Service

An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.


1 answer

Sort by: Most helpful
  1. Q&A Assist
    2026-04-02T18:55:50.6366667+00:00
    AI generated content. This question contains an answer created with AI from Q&A Assist. Learn more

    Per-model deployment metrics are exposed as dimensions on the Foundry resource metrics, not as separate child resources. The recommended pattern is to keep exporting metrics from the single Foundry resource and then slice those metrics by the model/deployment dimensions both in Metrics explorer and in Log Analytics.

    Concretely:

    1. Diagnostic settings remain at the Foundry resource level
      • Configure diagnostic settings on the Foundry resource only.
      • Under Metrics, select AllMetrics and send them to a Log Analytics workspace. No extra diagnostic setting is required per deployment.
    2. Use the Models metric namespace and dimensions in Metrics explorer
      In Azure Monitor > Metrics:
      1. Select Monitor in the Azure portal and then Metrics.
      2. Set the scope to the Foundry resource (or a resource group/subscription and filter Resource types to Foundry Tools).
      3. Choose metrics from the Models category (recommended over the legacy Azure OpenAI category). These metrics include dimensions such as ModelDeploymentName, ModelName, ModelVersion, and Region.
      4. Use Add splitting and split by ModelDeploymentName (or ModelName) to see a separate time series per deployment/model. This is the supported way to monitor each deployment individually.
    3. Query per-model metrics in Log Analytics using dimensions
      After metrics are exported via diagnostic settings, query them in Log Analytics:
      1. From the Foundry resource, go to Monitoring > Logs and switch to KQL mode.
      2. Start with the sample query to inspect the schema:
              AzureMetrics
              | take 100
              | project TimeGenerated, MetricName, Total, Count, Maximum, Minimum, Average, TimeGrain, UnitName
        
      3. Then filter and group by model/deployment dimensions exposed for the metric (for example, ModelDeploymentName, ModelName, Region). For instance, to aggregate ModelRequests per deployment over time, extend the query to summarize by those dimensions once they appear in the table.
      Metrics are still stored once per Foundry resource; the per-deployment view is obtained by grouping on these dimensions in KQL rather than by having separate Azure resources.
    4. Use resource logs for request-level per-deployment monitoring
      For detailed, request-level monitoring per deployment:
      • In diagnostic settings, enable the RequestResponse log category (and optionally Trace and Audit) and send them to the same workspace.
      • Query those logs in Log Analytics and filter by model/deployment fields to analyze latency, status codes, and usage per deployment.

    In summary, the recommended and supported approach is to keep a single diagnostic setting on the Foundry resource, export AllMetrics to Log Analytics, and then rely on the model/deployment dimensions in the Models metric category and in AzureMetrics/log tables to obtain per-model and per-deployment monitoring views.


    References:

    0 comments No comments

Your answer

Answers can be marked as 'Accepted' by the question author and 'Recommended' by moderators, which helps users know the answer solved the author's problem.