Speed up 'dlgradient' with parallelism?
4 次查看(过去 30 天)
显示 更早的评论
Hi all,
I am wondering if there is a way to speed up the 'dlgradient' function evaluation using parallelism or GPUs.
0 个评论
回答(1 个)
Jon Cherrie
2021-4-12
You can use a GPU for the dlgradient computation by using a gpuArray with dlarray.
In this example, the minibtachqueue, puts data on to the GPU and thus the GPU is used for the rest of the computation, both the "forward" pass the "backward" (gradient) pass:
1 个评论
Luis Hernandez
2023-11-14
Hello.
I've been trying to use the functions 'dlgradient' and 'dlfeval' with gpuArray inputs so that matlab will use my GPU. Unofrtunately, they only work when I pass dlarray inputs.
What is the workaround for this? what is the minibatch doing that allows you to work with gpuArray?
Thanks!
-L
另请参阅
Community Treasure Hunt
Find the treasures in MATLAB Central and discover how the community can help you!
Start Hunting!