53 views (last 30 days)

Show older comments

In my Matlab script, I have one line of code which uses the chain rule to calculate the derivative of a vector A (or an array) with respect to another vector B, it's something like:

dA/dB = dP/dB * dQ/dP * dA/dQ.

The 3 terms on the right hand side of the equation are all sparse (non-square non-diagonal) matrices, and in each matrix, the number of rows and of columns are both about 4000. This line of code does give correct derivatives, but the matrix multiplication is too slow! It's so slow that I can't really use this code..

In Matlab Workspace, the 3 matrices are shown to be "sparse double". Is it strange that sparse matrix multiplication takes so long in Matlab?

I would appreciate any suggestion, thank you!

-Paula

John D'Errico
on 24 Feb 2021

Edited: John D'Errico
on 24 Feb 2021

The answer is easy. It depends!

What does it depend on? How sparse are your matrices? A lot of people think if half the entries in a matrix are zero, then using sparse form is right. WRONG.You don't really gain much there. In fact, it probably runs MUCH more slowly. As a test on a relatvely small pair of matrices...

As = sprand(1000,1000,0.5); % 50% non-zero, so 500 non-zeros per row

Af = full(As);

Bs = sprand(1000,1000,0.5);

Bf = full(Bs);

timeit(@() As*Bs)

timeit(@() Af*Bf)

And that was only for a 1000x1000 pair of matrices.

A significant problem is that sparse matrix multiplication does not use multiple processors. On my own computer, I have 8 real cores. But that sparse multiply uses only ONE of them, whereas the full matrix multiply uses all 8 in parallel.

Now, let me try it again, but for larger matrices that are truly sparse.

As = sprand(5000,5000,0.001); % 0.1% non-zero, roughly 5 non-zeros per row.

Af = full(As);

Bs = sprand(5000,5000,0.001);

Bf = full(Bs);

timeit(@() As*Bs)

timeit(@() Af*Bf)

As you see in the second test, the sparse multiply whizzes past the full multiply. We are talking tortoise and hare here.

So the real problem is you most likely do not have matrices that are even remotely sparse. Hey, I have a lot of zeros. So sparse must be good. Just having a lot of zeros is not enough. A matrix needs to be seriously sparse for you to see a gain. But when that is the case, sparse is a HUGE benefit.

You may want to do some reading about sparse matrices and the concept of fill-in to understand why all of this works as it does.

For example, consider the matrices:

A1 = sprand(1000,1000,0.5);

B1 = sprand(1000,1000,0.5);

[numel(A1),nnz(A1),nnz(B1),nnz(A1*B1)]

A2 = sprand(5000,5000,0.001);

B2 = sprand(5000,5000,0.001);

[numel(A2),nnz(A2),nnz(B2),nnz(A2*B2)]

Do you see that the product A1*B1 is a completely full matrix? So even though A1 and B1 were sparse in theory, the product A1*B1 is completely full. It is still stored in sparse form of course. And if you then multiply that matrix by ANOTHER sparse matrix, the product will be even slower to compute.

John D'Errico
on 25 Feb 2021

Super. A question where you learn something is a good question!

Used properly, sparse matrices are an IMMENSELY valuable tool. I recall figuring out on one problem, before we had sparse matrices, that it would take days to solve a problem we needed to solve on the biggest, fastest computer we had available. With sparse matrices, it just happens in seconds. But your matrices need to be truly sparse to show a real gain.

Find the treasures in MATLAB Central and discover how the community can help you!

Start Hunting!