[Numpy-discussion] np.diag(np.dot(A, B))
mathieu at mblondel.org
Fri May 22 06:15:10 EDT 2015
Right now I am using np.sum(A * B.T, axis=1) for dense data and I have
implemented a Cython routine for sparse data.
I haven't benched np.sum(A * B.T, axis=1) vs. np.einsum("ij,ji->i", A, B)
yet since I am mostly interested in the sparse case right now.
When A and B are C-style and Fortran-style, the optimal algorithm should be
computing the inner products along the diagonal using BLAS.
If not, I guess this will need some benchmarking.
Another use for this is to compute the row-wise L2 norms: np.diagdot(A,
On Fri, May 22, 2015 at 5:58 PM, David Cournapeau <cournape at gmail.com>
> On Fri, May 22, 2015 at 5:39 PM, Mathieu Blondel <mathieu at mblondel.org>
>> I often need to compute the equivalent of
>> np.diag(np.dot(A, B)).
>> Computing np.dot(A, B) is highly inefficient if you only need the
>> diagonal entries. Two more efficient ways of computing the same thing are
>> np.sum(A * B.T, axis=1)
>> np.einsum("ij,ji->i", A, B).
>> The first can allocate quite a lot of temporary memory.
>> The second can be quite cryptic for someone not familiar with einsum.
>> I assume that einsum does not compute np.dot(A, B), but I haven't
>> Since this is is quite a recurrent pattern, I was wondering if it would
>> be worth adding a dedicated function to NumPy and SciPy's sparse module. A
>> possible name would be "diagdot". The best performance would be obtained
>> when A is C-style and B fortran-style.
> Does your implementation use BLAS, or is just a a wrapper around einsum ?
> NumPy-Discussion mailing list
> NumPy-Discussion at scipy.org
-------------- next part --------------
An HTML attachment was scrubbed...
More information about the NumPy-Discussion