Skip to content
Discussion options

You must be logged in to vote

NER models broadly rely on two kinds of features: token features and context features.

Token features are details of the labeled tokens themselves. This can be the whole token, like how "John" is likely to be a name and "the" isn't, or details - capitalized words are more likely to be names, words ending in "-son" are likely to be names.

Context features are drawn from the surrounding words. So a word after "Mr" or "Miss" is likely to be a name, a word after "the" is not very likely, "my name is" is a big hint, and so on.

Exactly what features are considered varies by model architecture or configuration, and how much weight each feature has, and how those weights interact, is what the mod…

Replies: 1 comment 5 replies

Comment options

You must be logged in to vote
5 replies
@polm
Comment options

@JessitaMS
Comment options

@polm
Comment options

@JessitaMS
Comment options

@polm
Comment options

Answer selected by adrianeboyd
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feat / ner Feature: Named Entity Recognizer
2 participants