Skip to content
Discussion options

You must be logged in to vote

I think your code looks reasonable but I don't think it will be possible to train a large language model end-to-end in combination with a GNN for around 20k nodes. Alternatives include

  1. using a more light-weight language embedding model, e.g., bag-of-words or averaging Word2Vec embeddings
  2. Pre-processing the intermediate and final outputs of BERT and using them as input features

Replies: 1 comment 11 replies

Comment options

You must be logged in to vote
11 replies
@rusty1s
Comment options

@icedpanda
Comment options

@rusty1s
Comment options

@icedpanda
Comment options

@rusty1s
Comment options

Answer selected by icedpanda
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants