gpt-4.1

Neural Network

GPT-4.1 is a flagship large language model based on the strengths of GPT-4, but with significant improvements in reasoning, context processing, and multimodal capabilities. It easily handles simultaneous analysis of multi-page documents, databases, and source code of large software projects.

Main

/

Models

/

gpt-4.1
32 768

Max answer length

(in tokens)

1 047 576

Context size

(in tokens)

2 $

Prompt cost

(per 1M tokens)

8 $

Answer cost

(per 1M tokens)

0 $

Image prompt

(per 1K tokens)

Overview
Providers
API
bothub
BotHub: Try GPT Chat for Freebot

Caps remaining: 0 CAPS
Providers gpt-4.1On Bothub, you can select your own providers for requests. If you haven't made a selection, we will automatically find suitable providers who can handle the size and parameters of your request.
Code example and API for gpt-4.1We offer full access to the OpenAI API through our service. All our endpoints fully comply with OpenAI endpoints and can be used both with plugins and when developing your own software through the SDK.Create API key
Javascript
Python
Curl
import OpenAI from 'openai';
const openai = new OpenAI({
  apiKey: '<your bothub access token>',
  baseURL: 'https://bothub.chat/api/v2/openai/v1'
});


// Sync - Text generation 

async function main() {
  const chatCompletion = await openai.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }],
    model: 'gpt-4.1',
  });
} 

// Async - Text generation 

async function main() {
  const stream = await openai.chat.completions.create({
    messages: [{ role: 'user', content: 'Say this is a test' }],
    model: 'gpt-4.1',
    stream: true
  });

  for await (const chunk of stream) {
    const part: string | null = chunk.choices[0].delta?.content ?? null;
  }
} 
main();
illustaration

How it works gpt-4.1?

GPT-4.1 from OpenAI functions as a next-generation language model that can process, understand and generate text, create code, generate ideas, analyze documents and communicate in a dialogue format. The main difference from previous versions is the increased context volume (up to 1 million tokens).