Running NER on tokenized data only: KeyError: "Parameter 'E' for model 'hashembed' has not been allocated yet." #10507
-
|
Hello, I am trying to only run Named Entity Recognition and my dataset already contains tokenized text which I need to use since I need this exact tokenization for the indices. I do not need to train the pipeline on my dataset, I only need to apply it for the results. I was trying to apply only the NER-pipeline on a custom But then I recieved the following error: After researching this "Parameter 'E' for model 'hashembed' has not been allocated yet.", I realized that I have to run But I recieve this error as well already when trying to apply the What am I missing here? I can find no other infos about this error. Thank you in advance! |
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 4 replies
-
|
This is kind of the same issue as #10508, just with slightly different (admittedly confusing) error messages. I'm not sure exactly what you're trying to do, but you probably want to keep the The |
Beta Was this translation helpful? Give feedback.
-
|
Hello This error is popping up when I am trying to train a 'ner' component, while keeping the tok2vec component frozen. Can I use spacy-alignment? If yes, how to go about it? Here is my base_config.cfg: Here is the error: |
Beta Was this translation helpful? Give feedback.
This is kind of the same issue as #10508, just with slightly different (admittedly confusing) error messages.
I'm not sure exactly what you're trying to do, but you probably want to keep the
nercomponent fromen_ner_bc5cdr_mdrather than excluding it and trying to replace it with new uninitialized/untrained one.The
Docconstruction is fine, but it's unlikely thaten_ner_bc5cdr_mdhas been trained on BPE/wordpiece-y tokens like##naso you might not see good results. Is there a particular reason that you're using this tokenization?