This repository was archived by the owner on Nov 17, 2023. It is now read-only.
Accuracy in STN drops rapidly during training #12082
Unanswered
rohitrango
asked this question in
Q&A
Replies: 3 comments
-
|
Hi @rohitrango , thanks for your issue. @ThomasDelteil as ML guru to help you out. Also please try to send this issue to https://discuss.mxnet.io/ as more discussion on Accuracy and performance will be there. @mxnet-label-bot could you please add [question, performance, python] here? |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
Thank you @lanking520, I will post it there as well. |
Beta Was this translation helpful? Give feedback.
0 replies
-
|
mark. I want know that why add tanh to the last layer make STN works well. |
Beta Was this translation helpful? Give feedback.
0 replies
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
Description
While training a Spatial Transformer Network on a modified MNIST dataset (40*40 images with rotated, scaled and translated images), the accuracy increases and then drops suddenly. I have tried different architectures, and the custom initialisation as well, but the problem persists. A baseline architecture without STNs gets ~96% val. accuracy on the dataset.
Environment info
Here is the output of `diagnose.py'
I'm using Python mxnet-cu92==1.2.1.post1
Training log:
Minimum reproducible example
This is the localization net that I'm using:
Here is the main model (an STN followed by a normal feedforward net):
And this is my training loop:
I have initialised the weights and biases of the last regression layer to get the identity transform.
Steps to reproduce
What have you tried to solve it?
The same problem happens every time. What should I do? Is it a problem with the STN or the localisation network?
Beta Was this translation helpful? Give feedback.
All reactions