[docs] 100 fold inaccuracy in time.clock() function in MacOS
havens at cs.sfu.ca
Sun Oct 3 00:39:08 CEST 2010
I am teaching Python to undergrads here at SFU. I discovered that the time.clock() function yields the correct processor time on Windows machines in seconds but in MacOS 10.6.4 it produces the time in divided by 100. So a full minute real time elapsed yields 0.60 seconds. This is repeatable it seems.
Surely the underlying C function is the same in Darwin as other unix implementations. Have you seen this issue before? Perhaps the docs for this function could be updated if this is expected behaviour.
regards, Bill Havens
Dr. William S. (Bill) Havens http://www.cs.sfu.ca/~havens
School of Computing Science http://www.cs.sfu.ca/
Simon Fraser University http://www.sfu.ca/
Burnaby, British Columbia, Canada V5A 1S6
phone: +1.604.518.1624 fax: +1.604.733.0711
More information about the docs