I've typed this into python shell:
>>> 0.1*0.1
0.010000000000000002
I expected that 0.1*0.1 is not 0.01, because I know that 0.1 in base 10 is periodic in base 2.
>>> len(str(0.1*0.1))
4
I expected to get 20 as I've seen 20 characters above. Why do I get 4?
>>> str(0.1*0.1)
'0.01'
Ok, this explains why I len
gives me 4, but why does str
return '0.01'
?
>>> repr(0.1*0.1)
'0.010000000000000002'
Why does str
round but repr
not? (I have read this answer, but I would like to know how they have decided when str
rounds a float and when it doesn't)
>>> str(0.01) == str(0.0100000000001)
False
>>> str(0.01) == str(0.01000000000001)
True
So it seems to be a problem with the accuracy of floats. I thought Python would use IEEE 754 single precicion floats. So I've checked it like this:
#include <stdint.h>
#include <stdio.h> // printf
union myUnion {
uint32_t i; // unsigned integer 32-bit type (on every machine)
float f; // a type you want to play with
};
int main() {
union myUnion testVar;
testVar.f = 0.01000000000001f;
printf("%f
", testVar.f);
testVar.f = 0.01000000000000002f;
printf("%f
", testVar.f);
testVar.f = 0.01f*0.01f;
printf("%f
", testVar.f);
}
I got:
0.010000
0.010000
0.000100
Python gives me:
>>> 0.01000000000001
0.010000000000009999
>>> 0.01000000000000002
0.010000000000000019
>>> 0.01*0.01
0.0001
Why does Python give me these results?
(I use Python 2.6.5. If you know of differences in the Python versions, I would also be interested in them.)
See Question&Answers more detail:
os 与恶龙缠斗过久,自身亦成为恶龙;凝视深渊过久,深渊将回以凝视…