BNN optimization.
Compare changes
- Danilo Ferreira de Lima authored
+ 7
− 7
@@ -63,7 +63,7 @@ class BNN(nn.Module):
@@ -63,7 +63,7 @@ class BNN(nn.Module):
@@ -82,8 +82,8 @@ class BNN(nn.Module):
@@ -82,8 +82,8 @@ class BNN(nn.Module):
# and the only regularization is to prevent the weights from becoming > 18 + 3 sqrt(var) ~= 50, making this a very loose regularization.
# An alternative would be to set the (alpha, beta) both to very low values, whichmakes the hyper prior become closer to the non-informative Jeffrey's prior.
# Using this alternative (ie: (0.1, 0.1) for the weights' hyper prior) leads to very large lambda and numerical issues with the fit.
# The likelihood noise level is controlled by sigma in the likelihood and it should be allowed to be very broad, but different
@@ -92,8 +92,8 @@ class BNN(nn.Module):
@@ -92,8 +92,8 @@ class BNN(nn.Module):
# Making both alpha and beta small makes the gamma distribution closer to the Jeffey's prior, which makes it non-informative
# Since, after standardization, we know to expect the variance to be of order (1), we can select also alpha and beta leading to high variance in this range
@@ -201,10 +201,10 @@ class BNNModel(RegressorMixin, BaseEstimator):
@@ -201,10 +201,10 @@ class BNNModel(RegressorMixin, BaseEstimator):