On 01/09/2019 07:49 AM, Paul Koning via cctalk wrote:

Understanding rounding errors is perhaps the most significant part of "numerical methods", a subdivision of computer science not as widely known as it should be. I remember learning of the work of a scientist at DEC whose work was all about this: making the DEC math libraries not only efficient but accurate to the last bit. Apparently this isn't anywhere near as common as it should be. And I wonder how many computer models are used for answering important questions where the answers are significantly affected by numerical errors. Do the authors of those models know about these considerations? Maybe. Do the users of those models know? Probably not. paul
A real problem on the IBM 360 and 370 was their floating point scheme. They saved exponent bits by making the exponent a power of 16, instead of 2. This meant that the result of any calculation could end up normalized with up to 3 most-significant zeros. That would reduce the precision of the number by up to 3 bits, or a factor of 8. Some iterative solutions compared small differences in successive calculations to decide when they had converged sufficiently to stop. These could either stop early, or run on for a long time trying to reach convergence.

IBM eventually had to offer IEEE floating point format on later machines.

Jon

Reply via email to