Skip to content
Discussion options

You must be logged in to vote

Hi @TimoSowa!

I am aware that this may be an effect of the models being trained on written text including puctuation.

Yeah, punctuation likely has a significant impact on the performance of the dependency parser.

Is there any way to make the original model "robust" against missing punctuation without using this kind of preprocessing (some model configuration, retraining etc.)?

Other that what you are already doing (restoring the punctuation in a preprocessing step) you could train with custom data or custom data augmentation. There's nothing in the configuration that would make up for for the lack of punctuation in the training data I'm afraid.

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by adrianeboyd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
lang / de German language data and models feat / parser Feature: Dependency Parser perf / accuracy Performance: accuracy
2 participants