You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to overfit the db_mv3 model on a single scanned A4 size image, with fairly straightforward annotation. (The default Paddle detector performs almost perfectly when I run inference on it with default settings, but the outputs are sentence-wise, whereas I need character/word level outputs). So I've tried labeling the image with character level annotations and got the following results.
From my observation of the training logs, the loss seems to be stuck at some point (either around 1.0 or 0.8) and does not drop any further. I was thinking maybe its stuck in a local minima or something, but there doesn't seem to be an option to use techniques like momentum, etc. So can someone please help me figure this out or give me tips to improve my training?
PS. I cannot increase the dataset size (or share it since its confidential, but you can imagine a fairly easy to detect A4 scan of a normal quality page). Any help is welcome. Thank you!
This discussion was converted from issue #13338 on July 26, 2024 01:20.
Heading
Bold
Italic
Quote
Code
Link
Numbered list
Unordered list
Task list
Attach files
Mention
Reference
Menu
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
问题描述 / Problem Description
I am trying to overfit the db_mv3 model on a single scanned A4 size image, with fairly straightforward annotation. (The default Paddle detector performs almost perfectly when I run inference on it with default settings, but the outputs are sentence-wise, whereas I need character/word level outputs). So I've tried labeling the image with character level annotations and got the following results.
Then I retrained with a mix word-level annotations and some character-level labels, that didn't help much either.
From my observation of the training logs, the loss seems to be stuck at some point (either around 1.0 or 0.8) and does not drop any further. I was thinking maybe its stuck in a local minima or something, but there doesn't seem to be an option to use techniques like momentum, etc. So can someone please help me figure this out or give me tips to improve my training?
PS. I cannot increase the dataset size (or share it since its confidential, but you can imagine a fairly easy to detect A4 scan of a normal quality page). Any help is welcome. Thank you!
Beta Was this translation helpful? Give feedback.
All reactions