Skip to content
Discussion options

You must be logged in to vote

And when I explicitly run command ! spacy project run evaluate on the same dev dataset that I have used for evaluation while training, it shows very little F1 Score, Precision, and Recall.

By "very little", are you referring to the +/- 85% F-scores that are being printed for the best threshold cutoff? Because it looks to me like your model is in fact properly trained (just compare it to the baseline). If there is a discrepancy with the numbers reported during training, could you paste the output log of that as well?

Replies: 2 comments 10 replies

Comment options

You must be logged in to vote
9 replies
@karndeepsingh
Comment options

@svlandeg
Comment options

@karndeepsingh
Comment options

@svlandeg
Comment options

@karndeepsingh
Comment options

Answer selected by svlandeg
Comment options

You must be logged in to vote
1 reply
@karndeepsingh
Comment options

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
usage General spaCy usage feat / ner Feature: Named Entity Recognizer feat / rel Feature: Relation Extractor
2 participants