Replies: 8 comments 3 replies
-
Easily done 👍 o1 and o3 models can now be selected in v1.0.5 (pre-release) I'm not sure o3-mini will be a good translator as these new models are optimised for reasoning and may not have as much focus on multilingual training data. o1 reportedly scores lower than 4o for tasks that involved creative/informal writing, and I suspect that's a bigger factor in translation than logic. It seems pretty good on my test subtitles though. There could also be hidden costs due to reasoning tokens that don't form part of the returned translations, according to https://openai.com/api/pricing/ - Output tokens include internal reasoning tokens generated by the model that are not visible in API responses. Apparently it cost $0.07 to translate my standard 266 line test subtitle... which does not increase the risk of me using up my API credit before it expires by much 😆 Worth a try anyway! Let us know how you find it. |
Beta Was this translation helpful? Give feedback.
-
Thanks. but I see now that on my account I only have access to o1 and o1
mini (not o3).
I tried it with o1/ o1 mini and I get these errors:
INFO: Translating scene number 1
INFO: Translating with model o1-mini, Using API Base: None
INFO: HTTP Request: POST https://api.openai.com/v1/chat/completions
"HTTP/1.1 400 Bad Request"
ERROR: Error translating scene 1: Unexpected error communicating with the
provider
ERROR: Unrecoverable error in TranslateSceneCommand
…On Mon, Feb 3, 2025 at 11:32 PM machinewrapped ***@***.***> wrote:
Easily done 👍 o1 and o3 models can now be selected in v1.0.5 (pre-release)
https://github.com/machinewrapped/gpt-subtrans/releases/tag/v1.0.5
I'm not sure o3-mini will be a good translator as these new models are
optimised for reasoning and may not have as much focus on multilingual
training data. o1 reportedly scores lower than 4o for tasks that involved
creative/informal writing, and I suspect that's a bigger factor in
translation than logic. It seems pretty good on my test subtitles though.
There could also be hidden costs due to reasoning tokens that don't form
part of the returned translations, according to
https://openai.com/api/pricing/ -
*Output tokens include internal reasoning tokens generated by the model
that are not visible in API responses.*
Apparently it cost $0.07 to translate my standard 266 line test
subtitle... which does not increase the risk of me using up my API credit
before it expires by much 😆
Worth a try anyway! Let us know how you find it.
—
Reply to this email directly, view it on GitHub
<#214 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AE2SID2ULKA7ZH3REBVTZCL2N7N7BAVCNFSM6AAAAABWJCWNK2VHI2DSMVQWIX3LMV43URDJONRXK43TNFXW4Q3PNVWWK3TUHMYTEMBUG44DKOI>
.
You are receiving this because you authored the thread.Message ID:
***@***.***
.com>
|
Beta Was this translation helpful? Give feedback.
-
Oh, surprising - o3-mini appeared straight away and worked without much fuss for me. I didn't test with o1. Not sure what to recommend - perhaps try again in a few days, in case API availability is being phased in? |
Beta Was this translation helpful? Give feedback.
-
Beta Was this translation helpful? Give feedback.
-
I guess it would be nice to have the reasoning_effort parameter (I guess thats really the only parameter it has, other than the new/old max_completion_tokens which includes both reasoning and completion) |
Beta Was this translation helpful? Give feedback.
-
Updated 1.0.5 again with a custom client for DeepSeek. It exposes max_tokens in the settings, since apparently if it isn't specified it default to 4096 (output tokens). which is half of what the models actually support! It also extracts reasoning_tokens from the response, and also the actual reasoning - fun to see the model thinking about OCR errors and how to make sure the translation suits the style of the film! Line #59: "2." – probably a scene number or a typo. Since subtitles don't usually have just numbers, maybe it's a misplaced line or part of a previous sentence. Maybe it's a misread of "二" (two), but in context, perhaps it's a card in a game, like "Two." But the next lines are about gambling ("梭" is "all-in" in poker). So maybe line #59 is "2." referring to a card, so translated as "Two." #66: "今晚可以財色兼收了" – "財色兼收" means to obtain both wealth and beauty. So "Tonight I'll get both wealth and women!" fits the character's boastful tone. I'm not sure if the reasoning actually results in better translations - the non-reasoning models are already pretty good at this sort of thing, but it's at least reassuring to see the model considering context. I'm getting a lot of API errors from DeepSeek today, which is less fun! I'll build 1.0.5 for Mac tomorrow and make it the official latest release. |
Beta Was this translation helpful? Give feedback.
-
Well, I lied about the Mac build (PyInstaller issues once again), but 1.0.5 is now the official latest version anyway. I also added support for Flash Thinking models, so now the 3 main reasoning models can all be used for translation. deepseek-reasoner is still the only one where you can inspect the reasoning, sadly! |
Beta Was this translation helpful? Give feedback.
-
I'm finding that Gemini 2.0 Flash Thinking only works with really small batch sizes - like under 20 lines, otherwise it tends to time out or something. Quite frustrating, because it works very well in Google AI Studio with much larger requests, and the translation reasoning is excellent. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
OpenAI o3-mini model is now out. It outperforms 4o while being cheaper and it would be nice to be supported by GPT Subtrans.
Beta Was this translation helpful? Give feedback.
All reactions