Textcat model has extremely low score on predictions - too little sample (19 categories, 5400 rows). #12362
-
Hi everyone, I work with 5400 rows of text and labels to train a model to recognise which type of a contractual provision the text refers to. The number of categories is 19, each with the following number of texts to train on. They are usually one sentence to one shorter paragraph long).
With certain simpler text samples, the score is pretty high, such as here:
Whereas with some more complex clause the prediction does not even make it to the top 5 (albeit this one is nearly identical to the ones it was trained on):
The training evaluation when creating a new clean model:
I apologise for the wall of data above. My question is: is this a common occurence for a dataset of this size? I understand that it's far to small for it to work properly, but I would expect more accurate guesses, especially given how structured legalese is. Many thanks! |
Beta Was this translation helpful? Give feedback.
Replies: 1 comment 1 reply
-
Oh wow, I thought the evaluations are ordered but they are not, hence why I missed the predictions actually being correct. Apologies for the useless topic then! Is there a way to order them from largest to lowest score? |
Beta Was this translation helpful? Give feedback.
Oh wow, I thought the evaluations are ordered but they are not, hence why I missed the predictions actually being correct. Apologies for the useless topic then!
Is there a way to order them from largest to lowest score?