Explore the OpenAI models available in the Giselle workspace. These models are categorized based on their primary strengths and use cases, reflecting OpenAI’s platform structure.

Quick Comparison

The following table summarizes the key features of the OpenAI models available in Giselle.

ModelsGenerate TextInput ImageWeb SearchReasoningContext WindowMax Output TokensPricing (Input/Output per 1M tokens)Availability
o4-mini✅ (High)200k tokens100k tokens1.10/1.10 / 4.40Pro
o3✅ (Highest)200k tokens100k tokens10.00/10.00 / 40.00Pro
o3-mini✅ (High)200k tokens100k tokens1.10/1.10 / 4.40Pro
gpt-4.11M tokens32k tokens2.00/2.00 / 8.00Pro
gpt-4.1-mini1M tokens32k tokens0.40/0.40 / 1.60Pro
gpt-4.1-nano1M tokens32k tokens0.10/0.10 / 0.40Free
gpt-4o128k tokens16k tokens2.50/2.50 / 10.00Pro
gpt-4o-mini128k tokens16k tokens0.15/0.15 / 0.60Pro
gpt-image-1UnknownN/A5.00/5.00 / 40.00Pro

Please note that some features listed (like specific API functionalities e.g., fine-tuning, batch processing, specific tool use like audio or transcription) may not be directly exposed or available within the Giselle interface even if supported by the underlying OpenAI model.

Reasoning Models

These o-series models excel at complex, multi-step tasks involving reasoning.

o4-mini

A faster, more affordable o-series reasoning model optimized for effective reasoning with efficient performance in coding and visual tasks. It offers a balance between speed, cost, and reasoning capabilities. It supports image inputs and shares the same large context window and output limits as o3.

  • Context Window: 200,000 tokens
  • Max Output Tokens: 100,000 tokens
  • Knowledge Cutoff: June 1, 2024
  • Inputs: Text, Image
  • Availability: Pro Plan

o3

OpenAI’s most powerful reasoning model, setting a high standard for math, science, coding, and visual reasoning tasks. It excels at technical writing, instruction-following, and analyzing text, code, and images in multi-step problems. It supports image inputs and has a large context window, making it ideal for deep analysis and complex workflows requiring meticulous reasoning and stability.

  • Context Window: 200,000 tokens
  • Max Output Tokens: 100,000 tokens
  • Knowledge Cutoff: June 1, 2024
  • Inputs: Text, Image
  • Availability: Pro Plan

o3-mini

A smaller, efficient reasoning model providing high intelligence comparable to older models like o1-mini in terms of cost and latency. It supports key developer features like Structured Outputs and function calling. It’s a text-only model focused on reasoning tasks requiring high token limits.

  • Context Window: 200,000 tokens
  • Max Output Tokens: 100,000 tokens
  • Knowledge Cutoff: October 1, 2023
  • Inputs: Text only
  • Availability: Pro Plan

Flagship Models

Versatile, high-intelligence models suitable for a wide range of complex tasks.

gpt-4.1

The flagship GPT-4.1 model excels at complex tasks and problem-solving across domains. It features significantly improved coding abilities, instruction following, and a massive ~1 million token context window. It supports text and image inputs, making it well-suited for deep analysis of large documents or codebases.

  • Context Window: 1,047,576 tokens
  • Max Output Tokens: 32,768 tokens
  • Knowledge Cutoff: June 1, 2024
  • Inputs: Text, Image
  • Availability: Pro Plan

gpt-4o

The versatile GPT-4o model (“o” for “omni”) provides comprehensive capabilities including advanced text generation, multimodal image input, and integrated web search (available as a tool). It supports structured outputs and function calling. With a 128k token context window, it is ideal for complex analytical tasks, multimodal understanding, and general-purpose advanced applications requiring up-to-date information.

  • Context Window: 128,000 tokens
  • Max Output Tokens: 16,384 tokens
  • Knowledge Cutoff: October 1, 2023
  • Inputs: Text, Image
  • Availability: Pro Plan

Cost-Optimized Models

Smaller, faster models that cost less to run, suitable for balanced performance or focused tasks.

gpt-4.1-mini

Provides a balance between intelligence, speed, and cost within the GPT-4.1 series. It inherits the ~1 million token context window and improved instruction following, making it an attractive model for many use cases requiring large context handling at a lower price point than the full GPT-4.1. Supports text and image inputs.

  • Context Window: 1,047,576 tokens
  • Max Output Tokens: 32,768 tokens
  • Knowledge Cutoff: June 1, 2024
  • Inputs: Text, Image
  • Availability: Pro Plan

gpt-4.1-nano

The fastest, most cost-effective GPT-4.1 model. It brings the ~1 million token context window and improved capabilities of the series to the most budget-sensitive tasks like classification, autocompletion, and information extraction. Supports text and image inputs.

  • Context Window: 1,047,576 tokens
  • Max Output Tokens: 32,768 tokens
  • Knowledge Cutoff: June 1, 2024
  • Inputs: Text, Image
  • Availability: Pro Plan

gpt-4o-mini

A fast, affordable small model (“o” for “omni”) optimized for focused tasks. Supports both text and image inputs with structured outputs and function calling capabilities. It is ideal for tasks requiring moderate intelligence and rapid responses where cost is a primary consideration, within a 128k token context window.

  • Context Window: 128,000 tokens
  • Max Output Tokens: 16,384 tokens
  • Knowledge Cutoff: October 1, 2023
  • Inputs: Text, Image
  • Availability: Free Plan

Image Generation Models

These models are specialized in generating high-quality images from text and image inputs.

gpt-image-1

OpenAI’s state-of-the-art image generation model. It is a natively multimodal language model that accepts both text and image inputs and produces image outputs. The model offers different quality levels (Low, Medium, High) and supports various image dimensions, allowing for flexible generation based on use case requirements.

  • Pricing: Input text: 5.00per1Mtokens,Inputimages:5.00 per 1M tokens, Input images: 10.00 per 1M tokens, Output images: $40.00 per 1M tokens
  • Quality Options: Low, Medium, High
  • Supported Dimensions: 1024x1024, 1024x1536, 1536x1024
  • Knowledge Cutoff: April 2025 (estimate based on release date)
  • Inputs: Text, Image
  • Outputs: Image
  • Availability: Pro Plan

Model Selection Guide

Guidelines for selecting the optimal OpenAI model within Giselle:

  • For balanced reasoning, speed, and cost (including images, 200k context): o4-mini (Pro)
  • For the most powerful reasoning and complex analysis (including images, 200k context): o3 (Pro)
  • For text-based reasoning tasks requiring very high token limits (200k context): o3-mini (Pro)
  • For flagship performance on complex tasks with very large context (1M tokens): gpt-4.1 (Pro)
  • For balanced performance with very large context (1M tokens) at lower cost: gpt-4.1-mini (Pro)
  • For the most cost-effective option with very large context (1M tokens): gpt-4.1-nano (Pro)
  • For comprehensive, high-intelligence tasks with multimodal needs and web search (128k context): gpt-4o (Pro)
  • For cost-effective, focused tasks (including multimodal, 128k context): gpt-4o-mini (Free)
  • For high-quality image generation from text or image inputs: gpt-image-1 (Pro)

Practices for Giselle

We recommend gpt-4.1 as a versatile primary model in Giselle for Pro users. It offers an excellent balance of capability, intelligence, and features (including web search via tool) across various tasks like business document creation, analysis, and research.

For tasks demanding the absolute highest level of reasoning or handling very large contexts (up to 200k tokens), consider o3 or o4-mini. o3 is the top choice for depth and stability, while o4-mini provides strong reasoning with better cost-efficiency. o3-mini is a good text-only alternative for high-token reasoning tasks.

The GPT-4.1 series (gpt-4.1, gpt-4.1-mini, gpt-4.1-nano) introduces a massive ~1 million token context window across all tiers, along with improved coding and instruction following.

  • Use gpt-4.1 for the most demanding tasks requiring the largest context.
  • gpt-4.1-mini offers a balance for large-context tasks at a lower cost.
  • gpt-4.1-nano is the most economical choice for leveraging the million-token context window.

For users on the Free plan or those prioritizing cost and speed for moderately complex tasks, gpt-4o-mini (128k context) and gpt-4.1-nano (1M context) are the recommended choices.

For image generation needs, gpt-image-1 provides high-quality results and supports both text and image inputs. The model offers different quality tiers to balance cost and detail based on specific requirements.

By combining these models in workflows, you can leverage their specific strengths. For example, use gpt-4o for initial research with web search, then pass the results to o3 for deep analysis, or use gpt-4.1 to analyze extremely large documents.

For detailed specifications and the full range of models offered directly by OpenAI, please check the Official OpenAI Documentation.