The Numerical Algorithms Group (NAG) has produced up to four times better performance with multicore optimisation for materials science and quantum Monte Carlo applications and reductions of up to 25 per cent in runtimes with I/O tuning for an ocean modelling application. These are the early results of NAG's distributed Computational Science and Engineering (dCSE) support programme for HECToR (UK's national supercomputing facility), which now consists of over 30 dedicated application optimisation projects complementing the traditional HPC user support provided by NAG.
In the first project to complete, a key materials science code, CASTEP, used by academic researchers and industry, was enhanced with band-parallelism to allow the code to scale to more than 1,000 cores. The speed of CASTEP on a fixed number of cores was also improved by up to four times on the original, representing a potential saving of around $3m of computing resources over the remainder of the HECToR service. The CASTEP project showed the collaborative nature of the dCSE programme, with the University of York undertaking the core development in conjunction with NAG HPC staff and the Science and Technology Facilities Council.
In another project, an ocean modelling application known as NEMO (Nucleus for European Modelling of the Ocean) underwent optimisation including I/O techniques and variable resolution approaches to run 25 per cent faster on relevant use cases. This represents a $600,000 saving in computing resources for that project with potentially multi-million dollar savings across all NEMO users. The 6 person-month project was performed by a collaboration of the National Oceanography Centre and the University of Edinburgh working with NAG HPC staff.
A third project optimised a quantum Monte Carlo code (CASINO) for better performance on multicore nodes by introducing shared memory techniques and hierarchical parallelism. This resulted in performance gains of up to four times on quad-core nodes and further performance gains from I/O optimisations for simulations using more than 10,000 cores. Following NAG's work, the scientists were able to run on 40,000 cores of the Jaguar Petaflops supercomputer at Oak Ridge National Laboratory. This 12 person-month dCSE project was undertaken by NAG HPC staff working with users at University College London, and is estimated to have saved the researchers around $1m in computing resources on HECToR.
'These three examples of HPC software projects show the real performance advantages – and cost savings – to researchers from enhancing applications to run optimally on the latest HPC machines,' said Andrew Jones, vice-president of HPC Consulting at NAG. 'Investment in application performance and algorithms appropriate to the computer architecture has now become critical for efficient use of HPC resources and users' time.'