Luca Morselli
2017-05-26 18:55:02 UTC
Hi to all, a theano newbie here.
I'm trying to training a MLP on MNIST dataset with mini-batch sgd (using
cpu) following deeplearning.net tutorial.
The default batch-size is set to 20 but when I launch the script after 2
epoch I'm out of ram.
I noticed that every time train_model(index) is called something is stored
in RAM. If I use all of training data this doesn't happen.
What I'm missing?
Thanks.
I'm trying to training a MLP on MNIST dataset with mini-batch sgd (using
cpu) following deeplearning.net tutorial.
The default batch-size is set to 20 but when I launch the script after 2
epoch I'm out of ram.
I noticed that every time train_model(index) is called something is stored
in RAM. If I use all of training data this doesn't happen.
What I'm missing?
Thanks.
--
---
You received this message because you are subscribed to the Google Groups "theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to theano-users+***@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.
---
You received this message because you are subscribed to the Google Groups "theano-users" group.
To unsubscribe from this group and stop receiving emails from it, send an email to theano-users+***@googlegroups.com.
For more options, visit https://groups.google.com/d/optout.