Aaron Snoswell
2017-06-12 11:52:50 UTC
I'm working through the the DeepLearning.net tutorials using Windows 64
bit, Python 3.6 and Theano installed through conda.
I was able to run the Classifying MNIST digits using Logistic Regression
<http://deeplearning.net/tutorial/logreg.html> demo fine, and got the same
results as listed in the tutorial, hitting 4 epochs/second (about double
the listed CPU performance in the tutorial). I then tried running the MLP
tutorial code <http://deeplearning.net/tutorial/mlp.html> (classify MNIST
digits using a simple MLP). During execution, the process gobbles up memory
continuously until I get a MemoryError and the python crashes. Watching the
task manager, I will occasionally see the memory usage drop - I assume this
is the garbage collector kicking in, but it happens rarely.
<Loading Image...
>
I've tried adjusting the MLP 'batch_size' parameter;
- With a value of 1000 (therefore n_train_batches == 50) the code runs
until the patience condition causes it to stop (no crash)
- With the default of 20 (n_train_batches == 2500) the code gets to
epoch 17 and crashes
- With a value of 10 (n_train_batches == 5000) I only get to epoch 3
before it crashes
Is this behavior expected with the hardware specs of the laptop I'm running
on? I've attached my DxDiag results here, but I've got 20GB of ram on this
machine.
Just trying to figure out if this crashing behavior is expected, or if I'm
seeing a memory leak of some sort.
Thanks.
bit, Python 3.6 and Theano installed through conda.
I was able to run the Classifying MNIST digits using Logistic Regression
<http://deeplearning.net/tutorial/logreg.html> demo fine, and got the same
results as listed in the tutorial, hitting 4 epochs/second (about double
the listed CPU performance in the tutorial). I then tried running the MLP
tutorial code <http://deeplearning.net/tutorial/mlp.html> (classify MNIST
digits using a simple MLP). During execution, the process gobbles up memory
continuously until I get a MemoryError and the python crashes. Watching the
task manager, I will occasionally see the memory usage drop - I assume this
is the garbage collector kicking in, but it happens rarely.
<Loading Image...
I've tried adjusting the MLP 'batch_size' parameter;
- With a value of 1000 (therefore n_train_batches == 50) the code runs
until the patience condition causes it to stop (no crash)
- With the default of 20 (n_train_batches == 2500) the code gets to
epoch 17 and crashes
- With a value of 10 (n_train_batches == 5000) I only get to epoch 3
before it crashes
Is this behavior expected with the hardware specs of the laptop I'm running
on? I've attached my DxDiag results here, but I've got 20GB of ram on this
machine.
Just trying to figure out if this crashing behavior is expected, or if I'm
seeing a memory leak of some sort.
Thanks.
--
---
You received this message because you are subscribed to the Google Groups "theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to theano-users+***@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
---
You received this message because you are subscribed to the Google Groups "theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to theano-users+***@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.