-
Notifications
You must be signed in to change notification settings - Fork 299
[memo] High memory consumption and the places of doubts #180
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Comments
Interesting :), I think the analysis in the future should also be extended to the following datasets: https://archive.ics.uci.edu/ml/datasets/covertype They proved tricky. |
FYI, when we use Optuna with a tiny model, we consume around only 150MB.
|
I tested the memory usage for the following datasets:
The details of the memory usage are the followings:
Note that KNN failed in Higgs and some trainings for each dataset were canceled because of memory out error. |
This is from #259 by @franchuterivera.
|
Check if we can use |
I write down the current memory usage as a memo just in case when we encounter memory leak issues in the future.
This post is based on the current implementation.
When we run a dataset with the size of 300B, AutoPytorch consumes ~1.5GB and the followings are the major source of the memory consumptions:
When we run a dataset with the size of 300MB (400,000 instances x 80 features) such as Albert, AutoPytorch consumes ~2.5GB and the followings are the major source of the memory consumptions:
All the information was obtained by:
and the logger which I set for the debugging. Note that I also added
time.sleep(0.5)
before and after the line of interest to eliminate the possibilities of the influences from other elements and checked each line in detail.The text was updated successfully, but these errors were encountered: