Why is MATLAB gpuArray sparse matrix multiplication so fast despite using double precision?

6 次查看(过去 30 天)
I am working with multiplication of a large sparse matrix with a dense matrix using gpuArray. On my GTX 1080, MATLAB's sparse matrix multiplication runs in 5.04ms (multiplication only timed with tic/toc)
tic
gpu_mmm = gpu_matrix * gpu_input;
mvm_time = toc;
. I also have a CUDA 10.2 implementation of sparse matrix multiplication using cuSPARSE, which runs the same sparse matrix multiplication in 7.25ms (timed with the Nvidia profiler). However, my CUDA implementation uses float32, while the MATLAB implementation only supports sparse matrices of type double. To my knowledge, GPUs are much faster with single precision calculations compared to double precision calculations, so I am wondering why MATLAB performs this calculation faster despite the difference in precision.
  2 个评论
Thomas Barrett
Thomas Barrett 2021-2-9
Hi @Di Xiao . Did you get to the bottom of this? I am doing the same thing (multiplying a complex sparse matrix by a regular dense complex matrix), using gpuArray, and lately I am wondering if I will see a speedup using cuSparse instead. What do you think, based on your experience with this?
Di Xiao
Di Xiao 2021-6-24
Sorry for the late response! I redid the timing experiment with a 2080Ti and using cuSPARSE was faster for me by around 2x.

请先登录,再进行评论。

回答(2 个)

Andrea Picciau
Andrea Picciau 2020-4-30
Hello there!
The correct way to time GPU operations is by using gputimeit.
mvm_time = gputimeit(@() gpu_matrix*gpu_input, 1);
or, in alternative
gpu = gpuDevice();
tic
gpu_mmm = gpu_matrix * gpu_input;
wait(gpu);
mvm_time = toc;
I suggest you try measuring your code like this...
  2 个评论
Di Xiao
Di Xiao 2020-4-30
Thanks. In this case I got the same timing using gputimeit(). I think this might be because I have the assignment after the matrix multiplication, so there is a wait to synchronize?
Andrea Picciau
Andrea Picciau 2020-5-1
GPU operations are executed asynchronously, which means most of the time control is returned to the user right after the operations are launched. Wait makes sure you're measuring the whole duration of the computation, and gputimeit does something similar under the hood.

请先登录,再进行评论。


Edric Ellis
Edric Ellis 2020-4-30
You should use gputime it to time operations on the GPU (although I'm not certain it will actually make a difference in this case). Behind the scenes, gpuArray is simply using the cuSPARSE routines in double-precision, so it should show basically the same performance...
  2 个评论
Di Xiao
Di Xiao 2020-4-30
Interesting to know the gpuArray is just using the cuSPARSE routines in double-precision. I'll look at my CUDA code to see if there's something going on there then. In this case my sparse matrix is real, while my dense matrix is complex. In my cuSPARSE routine I do one multiplication for the real and one for the complex - maybe that leads to the difference?
Joss Knight
Joss Knight 2020-5-2
If you are doing two separate multiplies rather than promoting the sparse array to complex and using the cusparseCgemm routine, then that is almost certainly where the difference comes from. MATLAB is also very efficient about memory allocation so the remaining discrepancies could be to do with the way you are managing memory.

请先登录,再进行评论。

类别

Help CenterFile Exchange 中查找有关 GPU Computing 的更多信息

Community Treasure Hunt

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!

Translated by