Choose a graphic card to train SegNet for deep learning
1 次查看(过去 30 天)
显示 更早的评论
I just started to learn semantic segmentation using SegNet. I'm in the stage to choose a graphic card between GTX 1060 6gb or GTX 1060 3gb or GTX 1050 Ti 4gb. I'm very budget tight so I won't consider anything beyond GTX 1060 6gb. I need suggestion whether GTX 1060 6gb is sufficient to train SegNet? My data base is small and it won't exceed 1gb of images. Or can I start from GTX 1050 Ti 4gb? I don't mind about slower training speed (I can leave my computer on overnight) but I'm a bit worried if 4gb is enough?
Let's say can I use GTX 1050 Ti 4GB for this Matlab tutorial
https://au.mathworks.com/help/vision/examples/semantic-segmentation-using-deep-learning.html
Thanks.
4 个评论
Joss Knight
2017-11-20
To answer the question on precision, all Deep Learning is single precision. This is also true of most image processing, although there are some notable exceptions (for instance imgaussfilt which by default computes in double precision for accuracy reasons).
回答(1 个)
Joss Knight
2017-11-19
The GeForce 10 series will all work fine, they just have different capabilities and constraints. On a cheaper GPU you may need to keep the training MiniBatchSize down in order to lower the stress on memory and compute power. I've seen quite a lot of problems with kernel timeouts and other issues posted to Answers on the 960, 970, 1050 and 1060, but I do think most of the problems can be dealt with that way.
0 个评论
另请参阅
类别
在 Help Center 和 File Exchange 中查找有关 Image Data Workflows 的更多信息
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!