An Azure service that provides access to OpenAI’s GPT-3 models with enterprise capabilities.
Per-model deployment metrics are exposed as dimensions on the Foundry resource metrics, not as separate child resources. The recommended pattern is to keep exporting metrics from the single Foundry resource and then slice those metrics by the model/deployment dimensions both in Metrics explorer and in Log Analytics.
Concretely:
- Diagnostic settings remain at the Foundry resource level
- Configure diagnostic settings on the Foundry resource only.
- Under Metrics, select AllMetrics and send them to a Log Analytics workspace. No extra diagnostic setting is required per deployment.
- Use the Models metric namespace and dimensions in Metrics explorer
In Azure Monitor > Metrics:- Select Monitor in the Azure portal and then Metrics.
- Set the scope to the Foundry resource (or a resource group/subscription and filter Resource types to Foundry Tools).
- Choose metrics from the Models category (recommended over the legacy Azure OpenAI category). These metrics include dimensions such as
ModelDeploymentName,ModelName,ModelVersion, andRegion. - Use Add splitting and split by
ModelDeploymentName(orModelName) to see a separate time series per deployment/model. This is the supported way to monitor each deployment individually.
- Query per-model metrics in Log Analytics using dimensions
After metrics are exported via diagnostic settings, query them in Log Analytics:- From the Foundry resource, go to Monitoring > Logs and switch to KQL mode.
- Start with the sample query to inspect the schema:
AzureMetrics | take 100 | project TimeGenerated, MetricName, Total, Count, Maximum, Minimum, Average, TimeGrain, UnitName - Then filter and group by model/deployment dimensions exposed for the metric (for example,
ModelDeploymentName,ModelName,Region). For instance, to aggregateModelRequestsper deployment over time, extend the query to summarize by those dimensions once they appear in the table.
- Use resource logs for request-level per-deployment monitoring
For detailed, request-level monitoring per deployment:- In diagnostic settings, enable the RequestResponse log category (and optionally Trace and Audit) and send them to the same workspace.
- Query those logs in Log Analytics and filter by model/deployment fields to analyze latency, status codes, and usage per deployment.
In summary, the recommended and supported approach is to keep a single diagnostic setting on the Foundry resource, export AllMetrics to Log Analytics, and then rely on the model/deployment dimensions in the Models metric category and in AzureMetrics/log tables to obtain per-model and per-deployment monitoring views.
References: