Skip to main content

o1 system instruction temperature

Understanding the Differences When Using OpenAI "o1" Models

If you’re integrating OpenAI’s REST API and plan to use the "o1" series models, it’s essential to be aware of some key differences compared to other models like the "gpt" series. This blog post highlights two notable changes and provides examples to help you adjust your implementation.

No "system" Role

The "o1" models only recognize the "assistant" and "model" roles in the conversation format. The "system" role, commonly used in the "gpt" series to define behavior, is not applicable when using these models. Instead, the functionality and behavior are implied by the model itself or managed through the prompt and conversation context.

No "temperature" Setting

Unlike other models where you can adjust the randomness of responses using the "temperature" setting, the "o1" models do not allow this parameter. This means you cannot modify the deterministic or creative nature of responses for these models. Instead, responses will have default behavior based on the model’s configuration.

These differences are detailed in OpenAI’s documentation, which you can find in the documentation .


Practical Examples Using "curl"

Here are two examples to illustrate how to call the API when using "o1" models.

Example 1: Basic Chat Request

curl https://api.openai.com/v1/chat/completions \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "o1-preview",
    "messages": [
      {"role": "assistant", "content": "What can I help you with today?"},
      {"role": "model", "content": "Can you summarize the latest news about AI?"}
    ]
  }'

In this example:

  • The "model" field specifies an "o1" model.

  • The conversation includes messages with only "assistant" and "model" roles.

Example 2: Chain of Thought Reasoning

curl https://api.openai.com/v1/chat/completions \
  -H "Authorization: Bearer YOUR_API_KEY" \
  -H "Content-Type: application/json" \
  -d '{
    "model": "o1-preview",
    "messages": [
      {"role": "assistant", "content": "Let’s solve a math problem together."},
      {"role": "model", "content": "Sure! What problem do you have in mind?"},
      {"role": "assistant", "content": "What is the sum of 23 and 45?"}
    ]
  }'

Popular posts from this blog

Undefined global vim

Defining vim as global outside of Neovim When developing plugins for Neovim, particularly in Lua, developers often encounter the "Undefined global vim" warning. This warning can be a nuisance and disrupt the development workflow. However, there is a straightforward solution to this problem by configuring the Lua Language Server Protocol (LSP) to recognize 'vim' as a global variable. Getting "Undefined global vim" warning when developing Neovim plugin While developing Neovim plugins using Lua, the Lua language server might not recognize the 'vim' namespace by default. This leads to warnings about 'vim' being an undefined global variable. These warnings are not just annoying but can also clutter the development environment with unnecessary alerts, potentially hiding other important warnings or errors. Defining vim as global in Lua LSP configuration to get rid of the warning To resolve the "Undefined global vi...

npm run build base-href

Using NPM to specify base-href When building an Angular application, people usually use "ng" and pass arguments to that invocation. Typically, when wanting to hard code "base-href" in "index.html", one will issue: ng build --base-href='https://ngx.rktmb.org/foo' I used to build my angular apps through Bamboo or Jenkins and they have a "npm" plugin. I got the habit to build the application with "npm run build" before deploying it. But the development team once asked me to set the "--base-href='https://ngx.rktmb.org/foo'" parameter. npm run build --base-href='https://ngx.rktmb.org/foo did not set the base href in indext.html After looking for a while, I found https://github.com/angular/angular-cli/issues/13560 where it says: You need to use −− to pass arguments to npm scripts. This did the job! The command to issue is then: npm run build -- --base-href='https://ngx.rktmb.org/foo...

CopilotChat GlobFile Configuration

CopilotChat GlobFile Configuration Want to feed multiple files into GitHub Copilot Chat from Neovim without listing each one manually? Let's add a tiny feature that does exactly that: a file glob that includes full file contents . In this post, we'll walk through what CopilotChat.nvim offers out of the box, why the missing piece matters, and how to implement a custom #file_glob:<pattern> function to include the contents of all files matching a glob. Using Copilot Chat with Neovim CopilotChat.nvim brings GitHub Copilot's chat right into your editing flow. No context switching, no browser hopping — just type your prompt in a Neovim buffer and let the AI help you refactor code, write tests, or explain tricky functions. You can open the chat (for example) with a command like :CopilotChat , then provide extra context using built-in functions. That “extra context” is where the magic really happens. Built-in functio...