Video 67: Comments on bad predictions of the model #690
Replies: 2 comments
-
Hi @slawomirwojtas , There could be several reasons your model isn't working. It's hard to tell without seeing the full model/training/data setup (the Google Colab link you shared isn't working). -- The order of your optimizations shouldn't matter too much (e.g. -- The learning rate is one of the most important values you can tune. But it depends on your data/model setup. Generally, values of 0.01 or 0.001 are good starting points but with a simple dataset/model, the learning rate can highly effect results. |
Beta Was this translation helpful? Give feedback.
-
I could give a simple answer to it. When you shuffle the data (aka train and test) you make it not serial (like Linear means straight lines that mean no upside down. (like so, it can't predict on the data because it gets nonlinear data. For this problem you can use Nonlinear regression methods. that can handle the nonlinear data. (@mrdbourke does my answer is correct?) For more information see this article - https://statisticsbyjim.com/regression/curve-fitting-linear-nonlinear-regression/ I hope it helps. I hope it helps understand the meaning of linearity. (In machine learning Linear means serial data not shuffled data) |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
Uh oh!
There was an error while loading. Please reload this page.
-
https://colab.research.google.com/drive/1sdLTBoyL5TONCta9UquLk3zi-Oq3R_De?usp=sharing Oh no, why the link, isn't working...
I am shocked how badly this model works with simple linear regression data.
The only change I've made is a little shuffling for train/test sets, I don't know if that way is conventional.
Also, I changed the order of actions regarding optimizer - it make more sense to me to clear gradients after taking a step. This way I have two blocks in the loop: one about the loss, the other about the optimizer. Can it cause any issues?
Anyways, 300 epochs and I had to decrease learning rate because the model was not converging at the default 0.01 - the loss had two values again and again:
Does it mean the learning ratio is too high? Event with smaller learning rate of 0.001 the quality of predictions is rather low. How come we had better results during the lessons?
Beta Was this translation helpful? Give feedback.
All reactions