-
-
Notifications
You must be signed in to change notification settings - Fork 2.7k
Description
I'm not sure if the guide and the actual code match up, especially about how much data the gpt-4-1106-preview and gpt-4-0125-preview models can handle. The guide says both models can deal with the same amount of data at once. But, looking at the code, it seems there's a difference in their settings.
BetterChatGPT/src/constants/chat.ts
Lines 50 to 51 in ecad41f
| 'gpt-4-1106-preview': 128000, | |
| 'gpt-4-0125-preview': 4096, |
Version / Description / Context
gpt-4-0125-preview
Description: The latest GPT-4 model intended to reduce cases of “laziness” where the model doesn’t complete a task. Returns a maximum of 4,096 output tokens.
Context window: 128,000 tokens
gpt-4-1106-preview
Description: GPT-4 Turbo model featuring improved instruction following, JSON mode, reproducible outputs, parallel function calling, and more. Returns a maximum of 4,096 output tokens. This preview model is not yet suited for production traffic.
Context window: 128,000 tokens
Reference:
https://platform.openai.com/docs/models/gpt-4-and-gpt-4-turbo