You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Hi and thanks for your excellent work on the PaddleOCR library.
I’m using PPStructureV3 (PaddleOCR v3.0) as part of a pipeline to process scanned documents that contain a mix of text and tables. In general, the pipeline performs very well. However, I’ve encountered a specific limitation:
When processing pages that contain slightly skewed tables (e.g., due to minor scanning angle errors), the layout reconstruction fails to extract the table structure correctly. This occurs even when setting:
I have confirmed that the text boxes are detected reasonably well, but the resulting table structure is often completely broken or missed when the entire page (or its content) is slightly rotated (~3–7 degrees).
My main questions are:
Is there any internal skew correction in the pipeline beyond the document-level orientation classifier (use_doc_orientation_classify)?
Is the layout detection model (e.g. RT-DETR-L) expected to be rotation-invariant to minor skew, or does it require explicitly deskewed input?
Are there any recommended preprocessing steps (rotation, warping, etc.) you advise prior to using PPStructureV3 for better layout recognition on slightly rotated scans?
Would fine-tuning the layout model on slightly skewed data improve its robustness, or is this already accounted for in the pretraining dataset?
I am open to contribute or test suggested workarounds if needed.
Thanks again for your time and for sharing such a powerful tool.
reacted with thumbs up emoji reacted with thumbs down emoji reacted with laugh emoji reacted with hooray emoji reacted with confused emoji reacted with heart emoji reacted with rocket emoji reacted with eyes emoji
Uh oh!
There was an error while loading. Please reload this page.
-
Hi and thanks for your excellent work on the PaddleOCR library.
I’m using PPStructureV3 (PaddleOCR v3.0) as part of a pipeline to process scanned documents that contain a mix of text and tables. In general, the pipeline performs very well. However, I’ve encountered a specific limitation:
When processing pages that contain slightly skewed tables (e.g., due to minor scanning angle errors), the layout reconstruction fails to extract the table structure correctly. This occurs even when setting:
I have confirmed that the text boxes are detected reasonably well, but the resulting table structure is often completely broken or missed when the entire page (or its content) is slightly rotated (~3–7 degrees).
My main questions are:
Is there any internal skew correction in the pipeline beyond the document-level orientation classifier (use_doc_orientation_classify)?
Is the layout detection model (e.g. RT-DETR-L) expected to be rotation-invariant to minor skew, or does it require explicitly deskewed input?
Are there any recommended preprocessing steps (rotation, warping, etc.) you advise prior to using PPStructureV3 for better layout recognition on slightly rotated scans?
Would fine-tuning the layout model on slightly skewed data improve its robustness, or is this already accounted for in the pretraining dataset?
I am open to contribute or test suggested workarounds if needed.
Thanks again for your time and for sharing such a powerful tool.
Beta Was this translation helpful? Give feedback.
All reactions