This work requires the solution of a computationally intensive machine learning problem known as non-negative matrix factorization (NMF). Ludmil Alexandrov developed the approach for detecting mutation signatures and the software (SigProfiler) while at the Sanger Institute and continues to build on this work with his team at the University of California, San Diego (UCSD). Together, NVIDIA and the Mutographs teams at UCSD and the Sanger Institute teamed up to use GPUs to accelerate this research.
“Research projects such as the Mutographs Grand Challenge are just that—grand challenges that push the boundary of what’s possible,” said Pete Clapham, leader of the Informatics Support Group at the Wellcome Sanger Institute. “NVIDIA DGX systems provide considerable acceleration that enables the Mutographs team to, not only meet the project’s computational demands, but to drive it even further, efficiently delivering previously impossible results.”
NVIDIA GPUs accelerate the scientific application by offloading the most time-consuming parts of the code. While the Sanger Institute saves cost and improves performance by running the computationally intensive work on GPUs, the rest of the application still runs on the CPU. From the researcher’s perspective, the overall application runs faster because it’s using the parallel processing power of the GPU to improve performance.
In the current project, researchers are studying DNA from the tumors of 5,000 patients with five cancer types: pancreas, kidney, colorectal, and two kinds of esophageal cancer. Five synthetic data matrices that mimic one type of real-world mutational profiles were used for estimating compute performance. An NVIDIA DGX-1 system runs the NMF algorithm against the five matrices, while the corresponding replicated CPU jobs are executed in docker containers on OpenStack virtual machines (VMs), specifically 60 cores in Intel Xeon Skylake Processors with 2.6 GHz and 697.3 GB of random-access memory (RAM).
The NVIDIA DGX-1 is an integrated system for AI featuring eight NVIDIA V100 Tensor Core GPUs that connect through NVIDIA NVLink, the NVIDIA high-performance GPU interconnect, in a hybrid cube-mesh network. Together with dual-socket Intel Xeon CPUs and four 100 Gb NVIDIA Mellanox® InfiniBand network interface cards, the DGX-1 delivers one petaFLOPS of AI power, for unprecedented training performance. The DGX-1 system software, powerful libraries, and NVLink network are tuned for scaling up deep learning across all eight V100 Tensor Core GPUs to provide a flexible, maximum performance platform for the development and deployment of AI applications in both production and research settings.