, and it possesses highly effective and useful options. The mannequin has quite a lot of parameters and choices you’ll be able to select from, which it’s a must to accurately choose to optimize GPT-5’s efficiency in your software space.
On this article, I’ll deep-dive into the totally different choices you have got when utilizing GPT-5, and show you how to select the optimum settings to make it work effectively in your use case. I’ll talk about the totally different enter modalities you should use, the accessible options GPT-5 has, comparable to instruments and file add, and I’ll talk about the parameters you’ll be able to set for the mannequin.
This text just isn’t sponsored by OpenAI, and is solely a abstract of my experiences from utilizing GPT-5, discussing how you should use the mannequin successfully.

Why you need to use GPT-5
GPT-5 is a really highly effective mannequin you’ll be able to make the most of for all kinds of duties. You’ll be able to, for instance, use it for a chatbot assistant or to extract essential metadata from paperwork. Nevertheless, GPT-5 additionally has plenty of totally different choices and settings, plenty of which you’ll be able to learn extra about in OpenAI’s information to GPT-5. I’ll talk about how you can navigate all of those choices and optimally make the most of GPT-5 in your use case.
Multimodal talents
GPT-5 is a multimodal mannequin, that means you’ll be able to enter textual content, photos, and audio, and the mannequin will output textual content. You too can combine totally different modalities within the enter, for instance, inputting a picture and a immediate asking concerning the picture, and obtain a response. Inputting textual content is, after all, anticipated from an LLM, however the potential to enter photos and audio may be very highly effective.
As I’ve mentioned in earlier articles, VLMs are extraordinarily highly effective for his or her potential to straight perceive photos, which normally works higher than performing OCR on a picture after which understanding the extracted textual content. The identical idea applies to audio as effectively. You’ll be able to, for instance, straight ship in an audio clip, and never solely analyze the phrases within the clip, but in addition the pitch, speaking pace, and so forth from the audio clip. Multimodal understanding merely permits you a deeper understanding of the information you’re analyzing.
Instruments
Instruments is one other highly effective characteristic you have got accessible. You’ll be able to outline instruments that the mannequin can make the most of throughout execution, which turns GPT-5 into an agent. An instance of a easy device is the get_weather() perform:
def get_weather(metropolis: str):
return "Sunny"
You’ll be able to then make your customized instruments accessible to your mannequin, together with an outline and the parameters in your perform:
instruments = [
{
"type": "function",
"name": "get_weather",
"description": "Get today's weather.",
"parameters": {
"type": "object",
"properties": {
"city": {
"type": "string",
"description": "The city you want the weather for",
},
},
"required": ["city"],
},
},
]
It’s essential to make sure detailed and descriptive data in your perform definitions, together with an outline of the perform and the parameters to make the most of the perform.
You’ll be able to outline plenty of instruments to make accessible to your mannequin, but it surely’s essential to recollect the core ideas for AI device definitions:
- Instruments are effectively described
- Instruments don’t overlap
- Make it apparent to the mannequin when to make use of the perform. Ambiguity makes device utilization ineffective
Parameters
There are three primary parameters you need to care about when utilizing GPT-5:
- Reasoning effort
- Verbosity
- Structured output
I’ll now describe the totally different parameters and how you can method deciding on them.
Reasoning effort
Reasoning effort is a parameter the place you choose from:
Minimal reasoning primarily makes GPT-5 a non-reasoning mannequin and must be used for less complicated duties, the place you want fast responses. You’ll be able to, for instance, use minimal reasoning effort in a chat software the place the questions are easy to reply, and the customers count on fast responses.
The harder your activity is, the extra reasoning you need to use, although you need to take into accout the fee and latency of utilizing extra reasoning. Reasoning counts as output tokens, and on the time of writing this text, 10 USD / million tokens for GPT-5.
I normally experiment with the mannequin, ranging from the bottom reasoning effort. If I discover the mannequin struggles to provide high-quality responses, I transfer up on the reasoning stage, first from minimal -> low. I then proceed to check the mannequin and see how effectively it performs. It is best to try to make use of the bottom reasoning effort with acceptable high quality.
You’ll be able to set the reasoning effort with:
consumer = OpenAI()
request_params = {
"mannequin" = "gpt-5",
"enter" = messages,
"reasoning": {"effort": "medium"}, # may be: minimal, low, medium, excessive
}
consumer.responses.create(**request_params)
Verbosity
Verbosity is one other essential configurable parameter, and you may select from:
Verbosity units what number of output tokens (excluding considering tokens right here) the mannequin ought to output. The default is medium verbosity, which OpenAI has additionally acknowledged is basically the setting used for his or her earlier fashions.
Suppose you need the mannequin to generate longer and extra detailed responses, you need to set verbosity to excessive. Nevertheless, I largely discover myself selecting between low and medium verbosity.
- For chat functions, medium verbosity is sweet as a result of a really concise mannequin could make the customers really feel the mannequin is much less useful (plenty of customers desire some extra particulars in responses).
- For extraction functions, nonetheless, the place you solely need to output particular data, such because the date from a doc, I set the verbosity to low. This helps make sure the mannequin solely responds with the output I would like (the date), with out offering extra reasoning and context.
You’ll be able to set the verbosity stage with:
consumer = OpenAI()
request_params = {
"mannequin" = "gpt-5",
"enter" = messages,
"textual content" = {"verbosity": "medium"}, # may be: low, medium, excessive
}
consumer.responses.create(**request_params)
Structured output
Structured output is a strong setting you should use to make sure GPT-5 responds in JSON format. That is once more helpful if you wish to extract particular datapoints, and no different textual content, such because the date from a doc. This ensures that the mannequin responds with a legitimate JSON object, which you’ll be able to then parse. All metadata extraction I do makes use of this structured output, as this can be very helpful for guaranteeing consistency. You should utilize structured output by including the “textual content” key within the request params to GPT-5, comparable to under.
consumer = OpenAI()
request_params = {
"mannequin" = "gpt-5",
"enter" = messages,
"textual content" = {"format": {"sort": "json_object"}},
}
consumer.responses.create(**request_params)
Make certain to say “JSON” in your immediate; if not, you’ll get an error for those who’re utilizing structured output.
File add
File add is one other highly effective characteristic accessible by means of GPT-5. I mentioned earlier the multimodal talents of the mannequin. Nevertheless, in some situations, it’s helpful to add a doc straight and have OpenAI parse the doc. For instance, for those who haven’t carried out OCR or extracted photos from a doc but, you’ll be able to as a substitute add the doc on to OpenAI and ask it questions. From expertise, importing recordsdata can be quick, and also you’ll normally get fast responses, largely relying on the hassle you ask for.
Should you want fast responses from paperwork and don’t have time to make use of OCR first, file add is a strong characteristic you should use.
Downsides of GPT-5
GPT-5 additionally has some downsides. The principle draw back I’ve observed throughout use is that OpenAI doesn’t share the considering tokens whenever you use the mannequin. You’ll be able to solely entry a abstract of the considering.
That is very restrictive in stay functions, as a result of if you wish to use greater reasoning efforts (medium or excessive), you can’t stream any data from GPT-5 to the person, whereas the mannequin is considering, making for a poor person expertise. The choice is then to make use of decrease reasoning efforts, which ends up in decrease high quality outputs. Different frontier mannequin suppliers, comparable to Anthropic and Gemini, each have accessible considering tokens.
There’s additionally been plenty of dialogue about how GPT-5 is much less artistic than its predecessors, although that is normally not an enormous downside with the functions I’m engaged on, since creativity normally isn’t a requirement for API utilization of GPT-5.
Conclusion
On this article, I’ve supplied an summary of GPT-5 with the totally different parameters and choices, and how you can most successfully make the most of the mannequin. If used proper, GPT-5 is a really highly effective mannequin, although it naturally additionally comes with some downsides, the primary one from my perspective being that OpenAI doesn’t share the reasoning tokens. At any time when engaged on LLM functions, I at all times suggest having backup fashions accessible from different frontier mannequin suppliers. This might, for instance, be having GPT-5 as the primary mannequin, but when it fails, you’ll be able to fall again to utilizing Gemini 2.5 Professional from Google.
👉 Discover me on socials:
📩 Subscribe to my e-newsletter
🧑💻 Get in contact
✍️ Medium
You too can learn my different articles:
















