This post was written by Nathan Whitehead
A few days ago a friend came to me with a question about floating point. Let me start by saying that my friend knows his stuff, he doesn’t ask stupid questions. So he had my attention. He was working on some biosciences simulation code and was getting answers of a different precision than he expected on the GPU and wanted to know what was up.
Even expert CUDA programmers don’t always know all the intricacies of floating point. It’s a tricky topic. Even my friend, who is so cool he wears sunglasses indoors, needed some help. If you look at the NVIDIA CUDA forums, questions and concerns about floating point come up regularly.        Getting a handle on how to effectively use floating point is obviously very important if you are doing numeric computations in CUDA.
In an attempt to help out, Alex and I have written a short whitepaper about floating point on NVIDIA GPUs.
In the paper we talk about various issues related to floating point in CUDA. You will learn:
- How the IEEE 754 standard fits in with NVIDIA GPUs
- How fused multiply-add improves accuracy
- There’s more than one way to compute a dot product (we present three)
- How to make sense of different numerical results between CPU and GPU