[Numpy-discussion] calculating the mean and variance of a large float vector
Bruce Southey
bsouthey@gmail....
Fri Jun 6 08:56:50 CDT 2008
Alan McIntyre wrote:
> On Thu, Jun 5, 2008 at 10:16 PM, Keith Goodman <kwgoodman@gmail.com> wrote:
>
>> How can that lead to instability? If the last half-million values are
>> small then they won't have a big impact on the mean even if they are
>> ignored. The variance is a mean too (of the squares), so it should be
>> stable too. Or am I, once again, missing the point?
>>
>
> No, I just didn't think about it long enough, and I shouldn't have
> tried to make an example off the cuff. ;) After thinking about it
> again, I think some loss of accuracy is probably the worst that can
> happen.
> _______________________________________________
> Numpy-discussion mailing list
> Numpy-discussion@scipy.org
> http://projects.scipy.org/mailman/listinfo/numpy-discussion
>
>
Any problems are going to mainly due to the distribution of numbers
especially if there are very small numbers and very large numbers. This
is mitigated by numerical precision and algorithm - my guess is that it
will take a rather extreme case to cause you any problems.
Python and NumPy are already using high numerical precision (may depend
on architecture) and NumPy defines 32-bit, 64-bit and 128-bit precision
if you want to go higher (or lower). This means that calculations are
rather insensitive to numbers used so typically there is no reason for
any concern (ignoring the old Pentium FDIV bug,
http://en.wikipedia.org/wiki/Pentium_FDIV_bug ).
The second issue is the algorithm where you need to balance performance
with precision. For simple calculations:
http://en.wikipedia.org/wiki/Algorithms_for_calculating_variance
Bruce
More information about the Numpy-discussion
mailing list