Replies: 1 comment
-
Hello @SNMS95,
No, it considers the gradients of value_fn.
Yes.
This won't solve the issue. As said in #1201, if the gradients you feed LBFGS do not match the gradients of the function you are trying to minimize, LBFGS cannot ensure that the direction it outputs is a descent direction and so the may linesearch fail. If you could define a function If I was you, I would try to see what goes wrong with the verbose option of the linesearch first (to see if indeed there is a problem with the update direction fed to the linesearch not being a descent direction). Then you may try lbfgs without the linesearch but with a small learning rate. Finally, you may consider other quasi-newton methods that may have been developed recently. |
Beta Was this translation helpful? Give feedback.
Uh oh!
There was an error while loading. Please reload this page.
-
I have a rugged design space and would like to clip the gradients to allow L-BFGS as the optimizer. The issue is that I don't know how to achieve this. The issue lies in the
value_fn
argument that L-BFGS needs. For exampleDoes the
value_fn
also consider clipped gradients? My understanding is that it does not. It get the grad and value at the starting point (which are clipped) but for further evaluations of the line-search, it sees a unclipped design space.Is my understanding correct ?
If so, Do I need to write a custom ad rule for the value_fn to properly clip the gradients?
Beta Was this translation helpful? Give feedback.
All reactions