@@ -816,7 +816,13 @@ class HalfNormal(PositiveContinuous):
816
816
rv_op = halfnormal
817
817
818
818
@classmethod
819
- def dist (cls , sigma = None , tau = None , * args , ** kwargs ):
819
+ def dist (
820
+ cls ,
821
+ sigma : Optional [DIST_PARAMETER_TYPES ] = None ,
822
+ tau : Optional [DIST_PARAMETER_TYPES ] = None ,
823
+ * args ,
824
+ ** kwargs ,
825
+ ):
820
826
tau , sigma = get_tau_sigma (tau = tau , sigma = sigma )
821
827
822
828
return super ().dist ([0.0 , sigma ], ** kwargs )
@@ -948,7 +954,14 @@ class Wald(PositiveContinuous):
948
954
rv_op = wald
949
955
950
956
@classmethod
951
- def dist (cls , mu = None , lam = None , phi = None , alpha = 0.0 , ** kwargs ):
957
+ def dist (
958
+ cls ,
959
+ mu : Optional [DIST_PARAMETER_TYPES ] = None ,
960
+ lam : Optional [DIST_PARAMETER_TYPES ] = None ,
961
+ phi : Optional [DIST_PARAMETER_TYPES ] = None ,
962
+ alpha : Optional [DIST_PARAMETER_TYPES ] = 0.0 ,
963
+ ** kwargs ,
964
+ ):
952
965
mu , lam , phi = cls .get_mu_lam_phi (mu , lam , phi )
953
966
alpha = pt .as_tensor_variable (floatX (alpha ))
954
967
mu = pt .as_tensor_variable (floatX (mu ))
@@ -1115,7 +1128,16 @@ class Beta(UnitContinuous):
1115
1128
rv_op = pytensor .tensor .random .beta
1116
1129
1117
1130
@classmethod
1118
- def dist (cls , alpha = None , beta = None , mu = None , sigma = None , nu = None , * args , ** kwargs ):
1131
+ def dist (
1132
+ cls ,
1133
+ alpha : Optional [DIST_PARAMETER_TYPES ] = None ,
1134
+ beta : Optional [DIST_PARAMETER_TYPES ] = None ,
1135
+ mu : Optional [DIST_PARAMETER_TYPES ] = None ,
1136
+ sigma : Optional [DIST_PARAMETER_TYPES ] = None ,
1137
+ nu : Optional [DIST_PARAMETER_TYPES ] = None ,
1138
+ * args ,
1139
+ ** kwargs ,
1140
+ ):
1119
1141
alpha , beta = cls .get_alpha_beta (alpha , beta , mu , sigma , nu )
1120
1142
alpha = pt .as_tensor_variable (floatX (alpha ))
1121
1143
beta = pt .as_tensor_variable (floatX (beta ))
@@ -1243,7 +1265,7 @@ class Kumaraswamy(UnitContinuous):
1243
1265
rv_op = kumaraswamy
1244
1266
1245
1267
@classmethod
1246
- def dist (cls , a , b , * args , ** kwargs ):
1268
+ def dist (cls , a : DIST_PARAMETER_TYPES , b : DIST_PARAMETER_TYPES , * args , ** kwargs ):
1247
1269
a = pt .as_tensor_variable (floatX (a ))
1248
1270
b = pt .as_tensor_variable (floatX (b ))
1249
1271
@@ -1329,7 +1351,7 @@ class Exponential(PositiveContinuous):
1329
1351
rv_op = exponential
1330
1352
1331
1353
@classmethod
1332
- def dist (cls , lam , * args , ** kwargs ):
1354
+ def dist (cls , lam : DIST_PARAMETER_TYPES , * args , ** kwargs ):
1333
1355
lam = pt .as_tensor_variable (floatX (lam ))
1334
1356
1335
1357
# PyTensor exponential op is parametrized in terms of mu (1/lam)
0 commit comments