In this tutorial, when explaining the weight file, it mentioned that if a convolutional layer has batch norm, then it will get 4 parameters from the weight file. The 4 parameters are: bn_weight, bn_bias, bn_running_mean, bn_running_var. But for batch normalization, why do we still need bn_weight and bn_bias? What kind of role do the bn_weight and bn_bias cast?
I am a freshman of DL so my question may seem to be dull...I will appreciate it if you could tell me the answer or where to find the answer.