[Numpy-discussion] Dot/inner products with broadcasting?
Jaakko Luttinen
jaakko.luttinen@aalto...
Thu Mar 14 06:54:06 CDT 2013
Answering to myself, this pull request seems to implement an inner
product with broadcasting (inner1d) and many other useful functions:
https://github.com/numpy/numpy/pull/2954/
-J
On 03/13/2013 04:21 PM, Jaakko Luttinen wrote:
> Hi!
>
> How can I compute dot product (or similar multiply&sum operations)
> efficiently so that broadcasting is utilized?
> For multi-dimensional arrays, NumPy's inner and dot functions do not
> match the leading axes and use broadcasting, but instead the result has
> first the leading axes of the first input array and then the leading
> axes of the second input array.
>
> For instance, I would like to compute the following inner-product:
> np.sum(A*B, axis=-1)
>
> But numpy.inner gives:
> A = np.random.randn(2,3,4)
> B = np.random.randn(3,4)
> np.inner(A,B).shape
> # -> (2, 3, 3) instead of (2, 3)
>
> Similarly for dot product, I would like to compute for instance:
> np.sum(A[...,:,:,np.newaxis]*B[...,np.newaxis,:,:], axis=-2)
>
> But numpy.dot gives:
> In [12]: A = np.random.randn(2,3,4); B = np.random.randn(2,4,5)
> In [13]: np.dot(A,B).shape
> # -> (2, 3, 2, 5) instead of (2, 3, 5)
>
> I could use einsum for these operations, but I'm not sure whether that's
> as efficient as using some BLAS-supported(?) dot products.
>
> I couldn't find any function which could perform this kind of
> operations. NumPy's functions seem to either flatten the input arrays
> (vdot, outer) or just use the axes of the input arrays separately (dot,
> inner, tensordot).
>
> Any help?
>
> Best regards,
> Jaakko
> _______________________________________________
> NumPy-Discussion mailing list
> NumPy-Discussion@scipy.org
> http://mail.scipy.org/mailman/listinfo/numpy-discussion
>
More information about the NumPy-Discussion
mailing list