Note
Access to this page requires authorization. You can try signing in or changing directories.
Access to this page requires authorization. You can try changing directories.
This article explains how to change the model version and settings in the prompt builder. The model version and settings can affect the performance and behavior of the generative AI model.
Model selection
You can change the model by selecting Model at the top of the prompt builder. The dropdown menu allows you to select from the generative AI models that generate answers to your custom prompt.
The default model is GPT-4.1 mini. You can also use the more powerful GPT-4.1 model or o3 model. The versions of these models might change.
In addition, we propose GPT-5 chat (preview) and GPT-5 reasoning (preview). Check availability of models by region in Feature availability by regions for prompts.
Using prompts in Power Apps or Power Automate consumes AI Builder credits, while using prompts in Microsoft Copilot Studio consumes messages. Learn more in AI Builder licensing and credit management.
Models details
GPT model | Licensing rules | Functionalities |
---|---|---|
GPT-4.1 mini (Default model) |
Charged at the Basic rate. Consumes messages in Microsoft Copilot Studio (more information: Message scenarios) and AI Builder credits in Power Apps and flows (more information: Power Platform Licensing Guide) |
Trained on data up to June 2024. Context allowed up to 128K tokens. |
GPT-4.1 | Charged at the Standard rate. Consumes messages in Microsoft Copilot Studio (more information: Message scenarios) and AI Builder credits in Power Apps and flows (more information: Power Platform Licensing Guide) |
Trained on data up to June 2024. Context allowed up to 128K tokens. |
o3 | Charged at the Premium rate. Consumes messages in Microsoft Copilot Studio (more information: Message scenarios) and AI Builder credits in Power Apps and flows (more information: Power Platform Licensing Guide) |
Trained on data up to June 2024. Context allowed up to 200K tokens. |
GPT-5 chat (preview) | Charged at the Standard rate. Consumes messages in Microsoft Copilot Studio (more information: Message scenarios) and AI Builder credits in Power Apps and flows (more information: Power Platform Licensing Guide) |
Trained on data up to Septembre 2024. Context allowed up to 400K tokens. |
GPT-5 reasoning (preview) | Charged at the Premium rate. Consumes messages in Microsoft Copilot Studio (more information: Message scenarios) and AI Builder credits in Power Apps and flows (more information: Power Platform Licensing Guide) |
Trained on data up to October 2024. Context allowed up to 400K tokens. |
GPT-4o mini and GPT-4o continue to be used in U.S. government regions. These models follow licensing rules and offer functionalities comparable to GPT-4.1 mini and GPT-4.1, respectively.
Choose between the models based on region availability, functionalities, and use cases. Check Feature availability by regions for prompts and the following Models comparison.
Models comparison
Factors | GPT-4.1 mini (default) | GPT-4.1 | o3 | GPT-5 chat (preview) | GPT-5 reasoning (preview) |
---|---|---|---|---|---|
Cost | Basic pricing, more cost-effective | Standard pricing | Premium pricing | Standard pricing | Premium pricing |
Performance | Good for most tasks | Superior for complex tasks | Trained for reasoning tasks | Highest scores in document understanding and response accuracy | Highest score in reasoning tasks like planning or complex analysis |
Speed | Faster processing | Might be slower due to complexity | Slower, as it reasons before responding | Enhanced, compared to GPT 4.1 | Slow, as it reasons before responding |
Use cases | Summarization, information tasks, image and document processing | Image and document processing, advanced content creation tasks | Data analysis and reasoning tasks, image and document processing | More advanced document processing, responses requiring high accuracy | Handle highest complexity of data analysis and reasoning tasks |
Decision narrative
When you need a cost-effective solution for moderately complex tasks, have limited computational resources, or require faster processing, choose GPT-4.1 Mini. It's ideal for projects with budget constraints and applications like customer support or efficient code analysis.
When you're dealing with highly complex, multimodal tasks that require superior performance and detailed analysis, choose GPT-4.1. It's the better choice for large-scale projects where accuracy and advanced capabilities are crucial. Another scenario where it's a better choice is when you have the budget and computational resources to support it. GPT-4.1 is also preferable for long-term projects that might grow in complexity over time.
For projects requiring advanced reasoning capabilities, the o3 model excels. It's suitable for scenarios that demand sophisticated problem-solving and critical thinking. The o3 model excels in environments where nuanced reasoning, complex decision-making, and detailed analysis are important.
GPT-5 chat and GPT-5 reasoning models are the most up-to-date model available, incorporating recent developments in AI about response accuracy, answers being more context aware, better document and image processing, or data analysis.
Model settings
You can access the setting panel by selecting ... > Settings at the top of the prompt builder. You can change the following settings:
- Temperature: Lower temperatures lead to predictable results, while higher temperatures allow more diverse or creative responses.
- Record retrieval: Number of records retrieved for your knowledge sources.
- Include links in the response: When selected, the response includes link citations for the retrieved records.
Temperature
The slider allows you to select the temperature of the generative AI model. It varies between 0 and 1, and guides the generative AI model about how much creativity (1) vs deterministic answer (0) it should provide.
Temperature is a parameter that controls the randomness of the output generated by the AI model. A lower temperature results in more predictable and conservative outputs. To compare, a higher temperature allows for more creativity and diversity in the responses. It’s a way to fine-tune the balance between randomness and determinism in the model’s output.
By default, the temperature is 0, as in previously created prompts.
Temperature | Functionality | Use in |
---|---|---|
0 | More predictable and conservative outputs. Responses are more consistent. |
Prompts that require high accuracy and less variability. |
1 | More creativity and diversity in the responses. More varied and sometimes more innovative responses. |
Prompts that create new out-of-the-box content |
Adjusting the temperature can influence the model’s output, but it doesn't guarantee a specific result. The AI’s responses are inherently probabilistic and can vary even with the same temperature setting.
Note
The temperature setting isn't available for the o1 model, so the slider is disabled when the o1 model is selected.
Model updates
- During July 2025, the following models were updated:
- GPT-4.1 mini model replaced GPT-4o mini model. Existing prompts previously running on GPT-4o mini now run on GPT-4.1 mini, using same Basic rates.
- GPT-4.1 model replaced GPT-4o and is in GA. Existing prompts previously running on GPT-4o now run on GPT-4.1, using same Standard rates.
- o3 model replaced o1 model and is in GA. Existing prompts previously running on o1 now run on o3, using same Premium rates.
- In August 2025, we added GPT-5 chat and GPT-5 reasoning to the model selector (preview).