@@ -959,7 +959,7 @@ def dist(
959
959
mu : Optional [DIST_PARAMETER_TYPES ] = None ,
960
960
lam : Optional [DIST_PARAMETER_TYPES ] = None ,
961
961
phi : Optional [DIST_PARAMETER_TYPES ] = None ,
962
- alpha = 0.0 ,
962
+ alpha : Optional [ DIST_PARAMETER_TYPES ] = 0.0 ,
963
963
** kwargs ,
964
964
):
965
965
mu , lam , phi = cls .get_mu_lam_phi (mu , lam , phi )
@@ -1128,7 +1128,16 @@ class Beta(UnitContinuous):
1128
1128
rv_op = pytensor .tensor .random .beta
1129
1129
1130
1130
@classmethod
1131
- def dist (cls , alpha = None , beta = None , mu = None , sigma = None , nu = None , * args , ** kwargs ):
1131
+ def dist (
1132
+ cls ,
1133
+ alpha : Optional [DIST_PARAMETER_TYPES ] = None ,
1134
+ beta : Optional [DIST_PARAMETER_TYPES ] = None ,
1135
+ mu : Optional [DIST_PARAMETER_TYPES ] = None ,
1136
+ sigma : Optional [DIST_PARAMETER_TYPES ] = None ,
1137
+ nu : Optional [DIST_PARAMETER_TYPES ] = None ,
1138
+ * args ,
1139
+ ** kwargs ,
1140
+ ):
1132
1141
alpha , beta = cls .get_alpha_beta (alpha , beta , mu , sigma , nu )
1133
1142
alpha = pt .as_tensor_variable (floatX (alpha ))
1134
1143
beta = pt .as_tensor_variable (floatX (beta ))
@@ -1256,7 +1265,7 @@ class Kumaraswamy(UnitContinuous):
1256
1265
rv_op = kumaraswamy
1257
1266
1258
1267
@classmethod
1259
- def dist (cls , a , b , * args , ** kwargs ):
1268
+ def dist (cls , a : DIST_PARAMETER_TYPES , b : DIST_PARAMETER_TYPES , * args , ** kwargs ):
1260
1269
a = pt .as_tensor_variable (floatX (a ))
1261
1270
b = pt .as_tensor_variable (floatX (b ))
1262
1271
@@ -1342,7 +1351,7 @@ class Exponential(PositiveContinuous):
1342
1351
rv_op = exponential
1343
1352
1344
1353
@classmethod
1345
- def dist (cls , lam , * args , ** kwargs ):
1354
+ def dist (cls , lam : DIST_PARAMETER_TYPES , * args , ** kwargs ):
1346
1355
lam = pt .as_tensor_variable (floatX (lam ))
1347
1356
1348
1357
# PyTensor exponential op is parametrized in terms of mu (1/lam)
0 commit comments