-
Notifications
You must be signed in to change notification settings - Fork 299
Updating the hyperparameter search space #154
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Updating the hyperparameter search space #154
Conversation
@@ -77,7 +77,7 @@ def get_hyperparameter_search_space(dataset_properties: Optional[Dict] = None, | |||
lr: Tuple[Tuple, float, bool] = ((1e-5, 1e-1), 1e-2, True), | |||
beta1: Tuple[Tuple, float] = ((0.85, 0.999), 0.9), | |||
beta2: Tuple[Tuple, float] = ((0.9, 0.9999), 0.9), | |||
weight_decay: Tuple[Tuple, float] = ((0.0, 0.1), 0.0) | |||
weight_decay: Tuple[Tuple, float, bool] = ((0.0, 0.1), 0.0, True) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
in the paper, I think weight decay is not on a log scale
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@ravinkohli You are right, that is why I also wrote adds log sampled weight decay. Common practice has it like that:
https://ml.informatik.uni-freiburg.de/papers/16-AUTOML-AutoNet.pdf
@@ -93,7 +93,7 @@ def get_hyperparameter_search_space(dataset_properties: Optional[Dict] = None, | |||
default_value=beta2[1]) | |||
|
|||
weight_decay = UniformFloatHyperparameter('weight_decay', lower=weight_decay[0][0], upper=weight_decay[0][1], | |||
default_value=weight_decay[1]) | |||
default_value=weight_decay[1], log=weight_decay[2]) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
here also
@@ -93,7 +93,7 @@ def get_hyperparameter_search_space(dataset_properties: Optional[Dict] = None, | |||
default_value=alpha[1]) | |||
|
|||
weight_decay = UniformFloatHyperparameter('weight_decay', lower=weight_decay[0][0], upper=weight_decay[0][1], | |||
default_value=weight_decay[1]) | |||
default_value=weight_decay[1], log=weight_decay[2]) |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
same here
@@ -72,7 +72,7 @@ def get_properties(dataset_properties: Optional[Dict[str, Any]] = None) -> Dict[ | |||
@staticmethod | |||
def get_hyperparameter_search_space(dataset_properties: Optional[Dict] = None, | |||
lr: Tuple[Tuple, float, bool] = ((1e-5, 1e-1), 1e-2, True), | |||
weight_decay: Tuple[Tuple, float] = ((0.0, 0.1), 0.0), | |||
weight_decay: Tuple[Tuple, float, bool] = ((0.0, 0.1), 0.0, True), |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
also here
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for this PR. We'll need these changes when we want to compare this to the paper version. I don't think the weight decay is on a log scale, other than that these changes look good
@ravinkohli I would say that common practice has it that you sample on the log scale for the l2 regularization term too as it is here: https://ml.informatik.uni-freiburg.de/papers/16-AUTOML-AutoNet.pdf |
Closing the pull request since I added it in the cocktails branch and we can later merge it into development. |
Given the recent experiments with the cocktails, I think the hyperparameter search space needs this update in the case when the architecture is not restricted. And this is a general update and not only related to the cocktails.
https://arxiv.org/abs/2006.13799
Matches the search space from the paper above and adds log sampled
weight decay
values.