Replies: 3 comments 4 replies
-
I presume you ran the notebook on Win/Linux? your output isn't abnormal, I'm getting the same, if that reassures you and you'd expect the same output from the book on Mac. As for the quality of the trajectory, it's just luck. You could change the seed and try to get something better. Even if weights were not properly loaded and failed silently, it would have been much worse. On top of that, these hparams |
Beta Was this translation helpful? Give feedback.
-
I've seen the same on my Mac.. the output I get looks similar to the gibberish you posted above @Monrother . Even with the dropout issue mentioned, this doesn't make sense because the pre-trained GPT-2 weights should result in much more coherent output regardless of dropout right? I've also seen calling |
Beta Was this translation helpful? Give feedback.
-
I originally missed this discussion, and thanks for chiming in here @casinca. I remember getting the same results on Google Colab (same as on my Mac). Like you said, it could be due to the MKL/dropout behavior in PyTorch, but during inference, that should have been disabled via It's still weird though that some machines produce this. I am really curious what components causes this. @Monrother and @natwille1 Do you encounter this discrepancy for |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I cloned the repo, and the output after I loaded the GPT 2 weights were different than the one shown in the book.
For "Every effort moves you", my local output is
Every effort moves you toward an equal share for each vote plus half. Inequality is often not an accurate representation of human worth; to know the
which I'm not sure what it is talking about.
And from the book, it is:
Every effort moves you toward finding an ideal new way to practice something!
What makes us want to be on top of that?
Which makes more sense.
Is this expected?
Beta Was this translation helpful? Give feedback.
All reactions