[SciPy-User] operations on large arrays
Vincent Davis
vincent@vincentdavis....
Sun Mar 7 00:05:03 CST 2010
I just figured out that I had a few arrays that where taking up a bunch of
the memory. That said I still wonder if there is a better way.
*Vincent Davis
720-301-3003 *
vincent@vincentdavis.net
my blog <http://vincentdavis.net> |
LinkedIn<http://www.linkedin.com/in/vincentdavis>
On Sat, Mar 6, 2010 at 10:22 PM, Vincent Davis <vincent@vincentdavis.net>wrote:
> I have arrays of 8-20 rows and 230,000 column, all the data is float64
> I what to be able to find the difference in the correlation matrix between
> arrays
> let A and B be of size (10, 230000)
> np.corrcoef(a)-np.corrcoef(b)
>
> I can't seem to do this with more than 10000 columns at a time because of
> memory limitations. (about 9GB usable to python)
> Is there a better way?
>
> I also have problem finding the column means which is surprising to me, I
> was not able to get the column means for 10000 columns, but I can computer
> the corrcoeff ?
> np.mean(a, axis=0)
>
> Do I just need to divide up the job or is there a better approach?
>
> Thanks
>
> *Vincent Davis
> 720-301-3003 *
> vincent@vincentdavis.net
> my blog <http://vincentdavis.net> | LinkedIn<http://www.linkedin.com/in/vincentdavis>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: http://mail.scipy.org/pipermail/scipy-user/attachments/20100306/748ef99d/attachment.html
More information about the SciPy-User
mailing list