o4-mini-high

Neural Network

ChatGPT o4 Mini High is a model that operates in an enhanced 'reasoning effort' mode to improve accuracy and depth of analysis. It is ideal for tasks requiring strict logical inferences and complex multi-step processes. It is particularly useful in programming, academic research, as well as corporate and professional applications.

Main

/

Models

/

o4-mini-high
100 000

Max answer length

(in tokens)

200 000

Context size

(in tokens)

1,1 $

Prompt cost

(per 1M tokens)

4,4 $

Answer cost

(per 1M tokens)

0 $

Image prompt

(per 1K tokens)

Overview
Providers
API
bothub
BotHub: Try GPT Chat for Freebot

Caps remaining: 0 CAPS
Providers o4-mini-highOn Bothub, you can select your own providers for requests. If you haven't made a selection, we will automatically find suitable providers who can handle the size and parameters of your request.
Code example and API for o4-mini-highWe offer full access to the OpenAI API through our service. All our endpoints fully comply with OpenAI endpoints and can be used both with plugins and when developing your own software through the SDK.Create API key
Javascript
Python
Curl
import OpenAI from 'openai';
const openai = new OpenAI({
  apiKey: '<your bothub access token>',
  baseURL: 'https://bothub.chat/api/v2/openai/v1'
});


// Sync - Text generation 

async function main() {
  const chatCompletion = await openai.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }],
    model: 'o4-mini-high',
  });
} 

// Async - Text generation 

async function main() {
  const stream = await openai.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }],
    model: 'o4-mini-high',
    stream: true
  });

  for await (const chunk of stream) {
    const part: string | null = chunk.choices[0].delta?.content ?? null;
  }
} 
main();
illustaration

How it works o4-mini-high?

The o4 Mini High model performs more subtle hypothesis testing, tool call deliberation, and verification of intermediate results through additional computational resources and output time. In 'high' mode, it can outperform standard o4-mini by 10-15% in accuracy, but with a 20-30% response delay.