There’s a new computational workhorse in town. For decades, general matrix-matrix multiply—known as GEMM in Basic Linear Algebra Subroutines (BLAS) libraries—has been a standard benchmark for computational performance.
Russian scientists from Lomonosov Moscow State University used an ordinary GPU-accelerated desktop computer to solve complex quantum mechanics equations in just 15 minutes that would typically take two to three days on a large CPU-only supercomputer.
Adam McLaughlin, PhD student at Georgia Tech shares how he is using NVIDIA Tesla GPUs for his research on Betweenness Centrality – a graph analytics algorithm that tracks the most important vertices within a network.
Daniel Ambrosi, Artist and Photographer, is using NVIDIA GPUs in the Amazon cloud and CUDA to create giant 2D-stitched HDR panoramas called “Dreamscapes.” Ambrosi applies a modified version of Google’s DeepDream neural net visualization code to his o