JVM out of memory when running parfor on cluster
3 次查看(过去 30 天)
显示 更早的评论
I'm running parfor on cluster from matlabpool open and getting JVM out of memory error heap size. The same parfor works perfect when I open matlabpool with 'local' profile and run it just on one machine.
MDCS is installed properly and passes validation.
For sure its not lack of memory on remote cluster PC - it has plenty of memory left and system is win7 64. Both PCs has 16G of memory,one is 8 core,another 4 core.
Java heap size is set to 2048MB, java.opts is set to 4096MB and heap size in java application is also set high. All those settings works well either for serial or for local parfor.
Krzysztof
0 个评论
采纳的回答
Thomas Ibbotson
2014-1-30
It's possible that the worker JVMs are not picking up the java.opts file. Workers have a different startup folder to the client MATLAB, so in order to get them to pick up the java.opts I would recommend putting it in the matlab/bin/<arch> folder.
You can check the java heap memory on the workers by running the following command with a pool open:
pctRunOnAll('r = java.lang.Runtime.getRuntime(); r.maxMemory / (1024*1024)')
The first result will be the memory on the client machine followed by the memory on each worker in MB.
更多回答(0 个)
另请参阅
类别
在 Help Center 和 File Exchange 中查找有关 Startup and Shutdown 的更多信息
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!