Skip to content

SGD Dev#106

Merged
maximilianreimer merged 49 commits intomainfrom
sgd_dev_pr
Oct 19, 2021
Merged

SGD Dev#106
maximilianreimer merged 49 commits intomainfrom
sgd_dev_pr

Conversation

@goktug97
Copy link
Collaborator

@goktug97 goktug97 commented Oct 8, 2021

No description provided.

rishan92 and others added 30 commits April 15, 2021 13:49
I added some TODO notes in the code that suggesting refactorings
- Replace string reward type config with enum
- add loss function to the config
- make reward range dynamic
- add test to check reward range for every reward type
- convert state values to normal python values from Tensors
- Negate full training reward
Steven-Adriaensen and others added 16 commits September 19, 2021 14:15
Fix a bug that causes validation loss to be computed even if it is not used.
During meta-train a random subset/prefix of MNIST is used as inner training set. In the old training set (now sgd_train_100instances_deprecated) this could cause issues at the end of the epoch, since the training dataset size was typically not a multiple of the mini-batch size and for some cases the modulo mini-batch was very small (1 or 2), in turn causing feature calculation failure (NaN). This update reduces the inner training dataset sizes to the nearest multiple of 32. As a consequence for any batch-size that is a multiple of 32, the modulo batch should never be smaller than 32.

(cherry picked from commit 907eda2)
@goktug97 goktug97 mentioned this pull request Oct 8, 2021
@maximilianreimer maximilianreimer merged commit 62fc8c4 into main Oct 19, 2021
@maximilianreimer maximilianreimer deleted the sgd_dev_pr branch October 19, 2021 13:48
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

5 participants