Replies: 2 comments 5 replies
-
|
Going for simpler equations with good fit goes a long way. I added a small parsimony to my search to encourage this (around a 10th of the lowest expected loss of the most complex equation). I found that unless you have specific domain knowledge and justification for complex operators, they tend to over fit and sticking to more simple operators generalise better. |
Beta Was this translation helpful? Give feedback.
-
|
Hi, thanks for the reply. I am now looking if we could consider the evaluation/performance of validation dataset during the training rather than after training. For example, in each iteration/epoch, we could use train dataset to fit the model, then we test performance on validation dataset. If the loss of validation dataset decrease, go and train next iteration/epoch. If not, stop the training to avoid overfit. Is there any ways that I could reproduce this thought? Thank you |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
The equations , generated by Symbolic Regression model, perform well on in sample dataset and have r-square about 0.1 . But they perform bad on out sample dataset. Is there any way that I could improve the generalization of the model. Currently, I have 2 ideas:
If anyone have any good ideas on improvement of model's generalization, please give me some hints please. Thank you.
Beta Was this translation helpful? Give feedback.
All reactions