Skip to content
Discussion options

You must be logged in to vote

If I understand your question correctly, the closest term for this in machine learning literature is "multi modal data". Most research is focused on the specific case of text + image features, but most techniques map to this case as well. It's also known as "multi-field data", but that's a less used term. It is a hard topic to search for.

Generally what you do in this case is create an embedding for each field of your document and combine those somehow (concatenation is the default approach) to create a representation. How the fields are vectorized is up to you - for fields like title, using the same methods as normal spaCy Docs is likely useful, but for other fields a categorical represe…

Replies: 1 comment

Comment options

You must be logged in to vote
0 replies
Answer selected by polm
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
feat / tok2vec Feature: Token-to-vector layer and pretraining
2 participants