Replies: 2 comments
-
@tink2123 any ideas? |
Beta Was this translation helpful? Give feedback.
0 replies
-
any ideas? |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Please provide the following information to quickly locate the problem
tools/train.py -c (path to my config file for RE training with LMv2)
Description:
Hello, I am training RE models with PPOCR. I have a custom dataset, which works since it has allowed me train a decent model with VI_LayoutXLM . When i try to train a model with the LMV2 architecture (default config file, only tweaked the paths to my custom dataset). The training starts for the first few iterations but crashes at the first evaluation. I was at first surprised by this since all the code around the evaluation and the post-processing class (in KIE) makes no distinction between the different architectures, so I expected that if the evaluation worked with VI_XLM, it should work with LMV.
I dug a bit in the code, and it seems like it's not a simple syntax issue as the final error line would have one believe: the model output that is fed (
preds
) to the post processing class does contain apred_relations
element in the output of VI_XLM, in my case a tensor of shapeshape=[4, 263169, 7, 2]
. On the other hand, the model output with LMV2 contains apred_relations
, but it is empty apart from its structure:'pred_relations': [[], [], [], []]
I am guessing this means there is a bug in the output of LMV2 when used for RE? Or at least in the way its RE pairings are outputted or at least processed. I have seen #8995 but it did not help me.
I would be grateful for any help, since in my case, RE performance would be much improved by the use of visual features (which LMV2 supports as opposed to VI_XML)
Beta Was this translation helpful? Give feedback.
All reactions