Loss remain constant for training #4546
Replies: 2 comments 7 replies
-
I think you first need to fix your issues with your node feature matrix. Currently, it looks like |
Beta Was this translation helpful? Give feedback.
-
Thank you @rusty1s! That fixes the constant loss problem. If I separate the dataset into 90% train, 5% validation, and 5% test randomly. The model will classify all the nodes into the largest label class. Do you have any suggestions? Is this underfitting? |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
My task is a node classification problem on a 60M nodes graph. 19M nodes are labeled, and 40M nodes are unlabeled. I created the dataset with a 15M nodes training mask, a 2M nodes validation mask, and a 2M nodes testing mask, out of 29M labeled nodes. Node labels are 243 countries(geolocation info). Following is the code creating the dataset. The node feature is 1 for every node.
Some dataset statistics:
Node lables count
The outgoing degree histogram is
I used two layers GraphSAGE model with NeighborLoaders for the training.
The training loss remains constant.
I have some guesses for the reasons.
I am desperate on this issue. Do I have any bug in my code causing the constant loss?
Thank you very much.
Beta Was this translation helpful? Give feedback.
All reactions