Skip to content

Clarification on Input/Output Length Parameters for gpt-4-1106-preview and gpt-4-0125-preview Models #533

@MBaltz

Description

@MBaltz

I'm not sure if the guide and the actual code match up, especially about how much data the gpt-4-1106-preview and gpt-4-0125-preview models can handle. The guide says both models can deal with the same amount of data at once. But, looking at the code, it seems there's a difference in their settings.

'gpt-4-1106-preview': 128000,
'gpt-4-0125-preview': 4096,


Version / Description / Context

gpt-4-0125-preview
Description: The latest GPT-4 model intended to reduce cases of “laziness” where the model doesn’t complete a task. Returns a maximum of 4,096 output tokens.
Context window: 128,000 tokens

gpt-4-1106-preview
Description: GPT-4 Turbo model featuring improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Returns a maximum of 4,096 output tokens. This preview model is not yet suited for production traffic.
Context window: 128,000 tokens

Reference:
https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions