Case Studies Archive - nAG https://nag.com/case-studies/ Robust, trusted numerical software and computational expertise. Thu, 22 May 2025 16:11:06 +0000 en-US hourly 1 https://wordpress.org/?v=6.8.2 https://nag.com/wp-content/uploads/2023/11/android-chrome-192x192-1-150x150.png Case Studies Archive - nAG https://nag.com/case-studies/ 32 32 Powering the SKA Telescopes: HPC for the Next Generation of Radio Astronomy https://nag.com/case-studies/ska-telescopes-hpc-next-gen-radio-astronomy/ Thu, 22 May 2025 15:10:03 +0000 https://nag.com/?post_type=case-studies&p=63083 Discover how HPC powers the SKA telescopes. nAG engineer Sean Stansill shares insights on data processing, MSv4, and scaling science with software.

The post Powering the SKA Telescopes: HPC for the Next Generation of Radio Astronomy appeared first on nAG.

]]>

We’ve helped pave the way for scalable, high-performance systems that will power radio telescopes for decades to come.
Sean Stansill, HPC Software Engineer, nAG

Overview

The SKA Observatory (SKAO) is an ambitious, €2 billion international endeavour to build the world’s largest radio telescope arrays, capable of transforming humanity’s understanding of the universe. Behind this monumental effort lies an equally groundbreaking challenge: processing petabytes of raw data every day into high-resolution astronomical images.

 

nAG’s High-Performance Computing (HPC) software engineer, Sean Stansill and the nAG HPC team, plays a pivotal role in making this possible. Their work ensures the seamless transformation of vast volumes of data into science-ready data products, supporting the SKAO’s mission to accelerate discovery in radio astronomy and gravitational wave science.

Image Copyright: SKAO

Challenge: Turning an Ambitious Concept into an Operational Reality

While the SKA telescopes have been envisioned since late 1980s, the computing infrastructure required to realise them only recently became feasible. Described as “software telescopes,” the SKA project does not rely on a single giant dish. Instead, it combines data from hundreds of smaller dishes and thousands of antennas using complex software pipelines. The computing demand is unprecedented — the two systems must sustain read/write speeds around 8 terabytes per second, non-stop.

The real bottleneck? I/O performance. Unlike embarrassingly parallel tasks, radio astronomy workflows require tightly synchronised data movement across many nodes. This choreography is dictated by the physics of how radio waves are processed. Efficiently managing this scale and complexity is a frontier problem in modern HPC.

Solution: Engineering Scalable, Future-Proof HPC Software Pipelines 

nAG HPC Services are playing a significant role in the Science Data Processing (SDP) phase where raw telescope data becomes scientifically useful. Sean Stansill, nAG HPC Engineer working on the SKAO project has contributed to:

  • Development and integration of MSv4, a next-generation data format poised to become the global standard for radio telescopes.
  • Enhancement of DP3, a calibration tool critical to ensuring scientific accuracy, now with distributed processing support for greater scalability.
  • Leading input on I/O architecture and performance optimisation during an international review consisting of 24 institutions.

He has also championed object storage — a mainstay in cloud computing — as a novel approach within HPC. Although unconventional in this domain, it offers the flexibility and scalability needed for SKAO’s evolving data landscape, reflecting a broader convergence of HPC and big data technologies.

Impact: Accelerating Science and Scaling the Future of Astronomy

Sean’s work ensures that SKA software runs not just effectively, but efficiently — squeezing maximum performance from existing infrastructure within a cost-conscious budget. One optimisation experiment revealed 4× speed improvements simply by doubling the node memory, a change with outsized performance benefits at minimal investment.

These advancements are not confined to code alone — they enable the SKAO to integrate data processing as a core part of the telescopes, freeing astronomers from manual workflows and fast-tracking discovery. From imaging black hole jets to verifying gravitational waves detected by LIGO, SKAO’s science goals will be made possible by the robust, scalable systems nAG engineers are building.

Image Copyright: SKAO Image Author: SKAO

Collaboration and Global Reach

This effort is deeply collaborative. Sean works not only within the SKAO team but also alongside counterparts at the US National Radio Astronomy Observatory (NRAO) and the South African Radio Astronomy Observatory (SARAO). Weekly working groups ensure technical alignment and smooth software integration across continents.

Looking Ahead

As the SKAO prepares for full operations towards the end of the decade, the innovations taking place today are setting the standard for scientific computing tomorrow. nAG’s work is paving the way for resilient, high-performance data systems that will serve the radio astronomy community for decades.

We’ve helped pave the way for scalable, high-performance systems that will power radio telescopes for decades to come.
Sean Stansill, HPC Software Engineer, nAG

Key Takeaways

  • 8 TB/s I/O is needed – the SKAO’s HPC systems must continuously sustain these rates.
  • Global standards: MSv4 is becoming the benchmark format for radio astronomy worldwide.
  • Innovative architecture: Adoption of object storage marks a shift in HPC design for data-intensive workloads.
  • Scientific impact: From studying the cosmic down to understanding galaxy evolution, the SKAO’s discoveries depend on the Science Data Processor software that nAG’s experts are helping to develop.
  • Cost-effective performance: Hardware tuning yielded up to 4× efficiency gains with minimal investment.

The post Powering the SKA Telescopes: HPC for the Next Generation of Radio Astronomy appeared first on nAG.

]]>
Migration of High Performance Computing Workloads for Mahindra https://nag.com/case-studies/migration-of-high-performance-computing-workloads-for-mahindra/ Thu, 03 Oct 2024 13:01:06 +0000 https://nag.com/?post_type=case-studies&p=49889 NAG worked with Mahindra and Google Cloud to enable the OpenFrontEnd Toolkit, adding customizations, providing key interventions and knowledge transfer to optimize Mahindra's HPC.

The post Migration of High Performance Computing Workloads for Mahindra appeared first on nAG.

]]>

nAG are renowned experts in High Performance Computing (HPC), providing experience in system architecture, implementation and engineering applications. This combination aligned perfectly to develop and build infrastructure to migrate Mahindra’s workloads to the Google Cloud Platform (GCP).

The Challenge

Mahindra runs multiple on-premise HPC systems for complex engineering workloads, needed to design and verify their automotive products. As part of Mahindra’s digital strategy to move resources to Google Cloud, these HPC systems were to be consolidated using GCP compute, integrating their existing workflows, whilst ensuring a production-level environment.

The Solution

Several HPC clusters were built within GCP Compute, making use of Open Front End and Cluster Toolkit for creation and administration of the systems. Existing filesystems were also integrated into the environment, along with connection to application servers, and customisations made to system images and the environment, all as needed by the engineering teams.

The Result

Mahindra now have a production HPC capability to run their workloads in GCP, improving the engineering simulation turnaround, with the advantages of being able to scale their resources on-demand and make use new technology sooner. In addition, a test system environment was established, enabling trials of new software, configurations, etc., without affecting production workloads.

“The Open Front End (OFE) was a key objective of our HPC transformation to Cloud as it allows us to manage our cluster(s) from an intuitive portal and enables us to scale and provision our cluster quickly. nAG worked with Mahindra and GCP to enable OFE for us, add customizations, provide key interventions & knowledge transfer so we can be self-sufficient & derive the expected benefits from our HPC.”

Gaurav Sarda, Business Engagement & Adoption Lead – Cloud, Mahindra

 

The post Migration of High Performance Computing Workloads for Mahindra appeared first on nAG.

]]>
Oxford Professor uses nAG components to aid the design of new magnetic materials https://nag.com/case-studies/oxford-professor-uses-nag-components-to-aid-the-design-of-new-magnetic-materials/ Fri, 07 Jul 2023 10:24:15 +0000 https://nag.com/?post_type=case-studies&p=3163 Magnetic materials come in many different forms. The most familiar are ferromagnets, such as iron and nickel, which become magnetized in an external magnetic field and retain their magnetization after the field has been switched off.

The post Oxford Professor uses <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG components to aid the design of new magnetic materials appeared first on nAG.

]]>

Oxford Professor uses nAG components to aid the design of new magnetic materials

Magnetic materials come in many different forms. The most familiar are ferromagnets, such as iron and nickel, which become magnetized in an external magnetic field and retain their magnetization after the field has been switched off. Ferromagnets are used in compasses, electronic recording devices, transformers, speakers, and many other everyday devices. Other, more subtle, forms of magnetism also exist and are particularly interesting because of their role in creating new states of matter with potentially useful properties, such as high temperature superconductivity or quantum information processing devices.

Magnetism arises from the existence of magnetic dipole moments, like tiny bar magnets, on some or all of the constituent atoms of the substance. The size of the magnetic moment is governed by the electronic state of the atom, which in turn depends on the local surroundings of the atom and the way the atom is bonded to its neighbours. Experiments that determine the electronic state of atoms in magnetic substances are therefore extremely valuable for understanding the origin of magnetic behaviour and for engineering materials with new or improved magnetic properties.

The original impetus for developing the software to analyze the experiments described here came from the availability of new experimental probes for studying magnetism. Neutron spectroscopy, in particular, has become a very powerful probe of the magnetic state of atoms thanks largely to advances in instrumentation at spallation neutron sources like the ISIS Facility at the Rutherford Appleton Laboratory. This technique makes it possible to obtain accurate measurements of level splittings in atoms caused by interactions with the crystalline environment. Detailed information on the many-electron states of atoms can be obtained from these level splittings and the corresponding spectral intensities.

Armed with the experimental data, Professor Andrew Boothroyd, at the University of Oxford Physics Department, needed the analytical tools to work out the state of the electrons and hence understand their magnetic behaviour. The creation of SPECTRE – the system designed to undertake the data analysis – and its subsequent evolution took place over many years as specific problems arose during the course of the research.

Having used numerical routines from nAG in previous projects, Professor Boothroyd turned to nAG to provide the mathematical code required by SPECTRE. SPECTRE uses the most up-to-date atomic models, which makes the results very accurate and quantitative. In brief, SPECTRE calculates the neutron spectra in terms of a small number of unknown parameters, and determines these parameters by least-squares fitting to the data. The core of the calculation is a series of matrix diagonalizations carried out by nAG routines designed to handle Hermitian matrices. The least-squares fitting is also performed by nAG routines. The lowest energy eigenfunctions found by the program are used to calculate other experimentally accessible physical properties such as the magnetic susceptibility and specific heat capacity.

SPECTRE will be made available to other magnetic materials groups in the hope that it will make it easier for scientists to interpret neutron spectroscopy data and thereby contribute towards the improvement of magnet materials.

Professor Andrew Boothroyd said “nAG’s reputation for accuracy, flexibility and robustness made it the first choice for the calculations in SPECTRE. I am delighted to be working with nAG to make the fruits of one person’s labour freely available to the whole scientific community.”

Mick Pont, Principal Technical Consultant at nAG commented “Software such as that developed by Professor Boothroyd is of fundamental importance in advancing human knowledge in the field of magnetism. We at nAG are very proud that our numerical components have proved useful in this endeavour. Since nAG was founded on collaborative principles we are particularly pleased that SPECTRE will be available for the use of others.”

For more information about the availability of SPECTRE please contact Professor Boothroyd.

Contact details:
Professor Andrew Boothroyd, Department of Physics, The University of Oxford
(a.boothroyd@physics.ox.ac.uk)
http://xray.physics.ox.ac.uk/Boothroyd/

The post Oxford Professor uses <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG components to aid the design of new magnetic materials appeared first on nAG.

]]>
PowerGen optimizes power plant performance using nAG’s Algorithms https://nag.com/case-studies/powergen-optimizes-power-plant-performance-using-nags-algorithms/ Fri, 07 Jul 2023 09:14:05 +0000 https://nag.com/?post_type=case-studies&p=3159 Faced with an increasingly competitive power supply market and stricter environmental targets, optimizing the performance of its power plants has become a major challenge for PowerGen, a global generator, distributor and supplier of electricity.

The post PowerGen optimizes power plant performance using <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG’s Algorithms appeared first on nAG.

]]>

The Challenge

Faced with an increasingly competitive power supply market and stricter environmental targets, optimizing the performance of its power plants has become a major challenge for PowerGen, a global generator, distributor and supplier of electricity.

Playing a major part in achieving optimization at power stations is Dr Ian Mayes, Senior Engineer in the Software Engineering Group at PowerGen’s Power Technology Centre in Nottingham. With a background in physics, Dr Mayes develops mathematical models of engineering processes and, in particular, the use of these models to optimize performance.

Commenting Dr Mayes said: “In view of today’s commercial and environmental pressures, the criteria for optimization in a power plant are typically based on either minimizing the NOx emissions whilst limiting the amount of unburnt fuel (carbon) left in the boiler ash or minimizing the amount of unburnt fuel whilst setting a limit on the NOx emissions.”

Optimization of power plant performance based on these criteria represents a major challenge since it is very difficult to reduce both unburnt carbon and NOx at the same time.

The Solution

The combustion process within a coal-fired power plant is complex and subject to a number of variable parameters that can affect performance. These include the distribution of coal in the boiler, the amount and distribution of the air in the boiler and, sometimes the particle size of the ground coal, which can be sourced from several different coal mines.

There are constraints, however, on which parameters can be adjusted to achieve optimization. Fuel flow, for example, needs to be maintained in order to keep the amount of power generated constant. If this constraint is removed, the model would simply shut the plant down – great for reducing NOx emissions but not appropriate for generating power or profitability.

There are also constraints on steam temperatures which can limit the range of operation and also constraints on engineering the upper and lower boundaries of the control variables.

With such a complex process, in which both dependent and independent variables can play an important part in overall performance, it is clear that optimization represents a genuine challenge. And, it is a challenge that needs to be tackled, if the balance between commercial and environmental targets is to be achieved.

There was really only one effective way forward – the development of a software model of the combustion process in the boiler of the coal-fired power plant that can account for all these variables.

Dr Mayes explained: “With constantly changing environmental legislation and commercial pressures, there is a need for continuous optmization of the power station boiler and traditional combustion testing only provided a snap-shot. It was also too expensive to frequently run tests on the actual plant – it simply wasn’t feasible. So, we’ve developed mathematical models to enable the process to be optimized based on certain criteria – a general concept that can be applied across many different industrial systems. The key feature is the need for a model of the process that allows us to use optimization techniques to find better solutions.”

In writing and developing the model, PowerGen turned to nAG for mathematical analysis software that could be integrated within its specialist application.

Dr Mayes explained: “With so many variables and parameters that can influence performance, a large degree of number crunching is required within our model. Whilst we concentrated on developing the specialist software for our application, we used the nAG Fortran Library to carry out mathematical manipulation of data within the model. After all, there is little point in reinventing the wheel. We looked at what was available and nAG offered precisely what we required.

Not only that, nAG has a reputation for high quality, tried and tested software that is well documented and supported by first class technical help, if required. In fact, whenever a mathematical operation of any complexity is required, we check to see whether there are nAG routines that can do the job.”

The Results

Today, PowerGen has developed boiler modelling software, incorporating nAG algorithms, that enables them to optimize the performance of their power plants based on specific criteria.

If legislation and commercial pressures change, then the criteria for optimum performance will change and parameters have to be altered. Importantly, the model enables PowerGen to determine the effects on performance of these changes, before applying them to the actual power plant. Tests can be run quickly and cost effectively, so that the parameters can be set to achieve optimum performance based on specific criteria.

Discussing the benefits of nAG and the contribution the company has made to PowerGen, Dr Mayes said: “There is no doubt that by using nAG products our development time for software applications involving complex mathematics has been reduced significantly. There is also a great comfort factor in using nAG software. We know their products are accurate, reliable and robust, and they will not fall down.”

The nAG Fortran Library

The nAG Fortran Library contains over 1,700 complex and highly sophisticated user callable routines for mathematical and statistical computation, which many organisations integrate with a variety of applications including Visual Basic, VBA, Excel, Fortran and C/C++ programs.

The routines cover the following areas: Eigenvalues and Eigenvectors; FFTs; Interpolation; Linear Algebra; Optimization; Partial and Ordinary Differential Equations; Quadrature, Curve and Surface Fitting; Random Number Generation; and, Statistics.

The correctness of each Library routine is evaluated and verified by specifically written test programs that are performed on each of the machine ranges for which the Library is available, and only when an implementation satisfies nAG’s stringent accuracy standards is it released.

Also available in C, Fortran 90 and high-performance computing versions, nAG’s algorithms are backed-up by an extensive range of services and support facilities including customization. Underpinning the quality of all nAG software is our renowned and comprehensive documentation.

Dr Ian Mayes Senior Engineer
Software Engineering Group
PowerGen Power Technology Centre Nottingham

The post PowerGen optimizes power plant performance using <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG’s Algorithms appeared first on nAG.

]]>
Load balancing issues uncovered in a particle tracking application https://nag.com/case-studies/load-balancing-issues-uncovered-in-a-particle-tracking-application/ Thu, 06 Jul 2023 16:19:33 +0000 https://nag.com/?post_type=case-studies&p=3152 The Centre for Environment, Fisheries and Aquaculture Science (Cefas) use an off-line particle tracking model that requires velocity fields from a hydrodynamic model.

The post Load balancing issues uncovered in a particle tracking application appeared first on nAG.

]]>

Project Background

The Centre for Environment, Fisheries and Aquaculture Science (Cefas) is a world leader in marine science and technology, providing innovative solutions for the aquatic environment, biodiversity and food security. The centre uses behaviour and transport models to address a range of marine management questions.

They use an off-line particle tracking model that requires velocity fields from a hydrodynamic model, which is known as the Individual Behaviour Model (IBM) GITM (General Individuals Transport Model) code. It includes physical particle advection and diffusion, and biological development and behaviour. The code is implemented in Fortran 90. It was originally sequential then parallelized with OpenMP.

Objective

The objective of this work was to investigate how to improve the parallel performance of the GITM application using OpenMP and to quantify the effort for it to run on distributed memory HPC systems using MPI.

Method

Audit – The GITM audit took two months. A set of performance metrics were used to assess the quality of performance and identify any limiting issues, these metrics relate to computational scaling and load balance. The audit identified that the application could further benefit from using multiple threads. The underlying cause was found to be related to load balance caused by inefficient cache usage and inefficient array alignments for vectorization. As the code was parallelized using OpenMP, the I/O using NetCDF was done sequentially.

Example of load balance test report

Report – Recommendations were made to improve the vectorization and computational performance of the application. Specifically, a change of compiler would assist in vectorization of masked Fortran 90 array operations. Improved array alignment would also enable more efficient vectorization and a reduction in the use of floating point division would enable more efficient computation. A long term recommendation was made for the use of MPI so that the I/O can be performed via the parallel NetCDF Library, rather than sequential access methods, to make it more scalable.

Code improvements  The code improvements are being carried out by the existing development with advice available from nAG when requested.

On-going – Work is currently progressing to improve I/O and vectorize further parts of the code.

Long term recommendation of the use of MPI is being considered so that the parallel NetCDF Library could be used for parallel I/O.

Note: This work was carried out by nAG staff working under the remit of the Performance Optimisation and Productivity Centre of Excellence in Computing Applications (POP)

The post Load balancing issues uncovered in a particle tracking application appeared first on nAG.

]]>
3 x Speed Improvement for Computational Fluid Dynamics Solver https://nag.com/case-studies/3-x-speed-improvement-for-computational-fluid-dynamics-solver/ Thu, 06 Jul 2023 15:47:19 +0000 https://nag.com/?post_type=case-studies&p=3146 zCFD by Zenotech is a computational fluid dynamics (CFD) solver for steady-state or time-dependent flow simulation. It is written in Python and C++ and parallelised with OpenMP and MPI.

The post 3 x Speed Improvement for Computational Fluid Dynamics Solver appeared first on nAG.

]]>

Real-world success stories of nAG HPC Consulting

zCFD by Zenotech is a computational fluid dynamics (CFD) solver for steady-state or time-dependent flow simulation. It is written in Python and C++ and parallelised with OpenMP and MPI.

nAG, acting through our role in the POP project, conducted a Performance Audit to identify potential areas for improvement. This identified that the code was spending a surprisingly large amount of time executing in serial and that one particular OpenMP loop was suffering from load imbalance. The POP team also noted that the CPU frequency was being lowered when the code was run on the maximum number of threads (12 for the machine used in the Audit). As a result of the nAG analysis, a number of changes were made to the code which resulted in significant improvements including:

  • Code ran 1.65x faster on 12 threads
  • A 3x performance improvement over the old code on 12 threads was observed when the modified code was applied to a test case 100x larger
  • The average cycle time fell from 3,253ms to 1,185ms, which corresponds to going from 10.4 GFLOPS to 30.6 GFLOPS for a single Broadwell socket
  • Load balancing was improved

The post 3 x Speed Improvement for Computational Fluid Dynamics Solver appeared first on nAG.

]]>
Audit of a Density Functional Theory application suggested a potentially improved communications scheme https://nag.com/case-studies/audit-of-a-density-functional-theory-application-suggested-a-potentially-improved-communications-scheme/ Thu, 06 Jul 2023 14:38:29 +0000 https://nag.com/?post_type=case-studies&p=3131 The objective of this work was to investigate the possibilities of improving the DFTB application performance on large systems distributed-memory machines.

The post Audit of a Density Functional Theory application suggested a potentially improved communications scheme appeared first on nAG.

]]>

Project Background

Software for Chemistry & Materials (SCM) is an Amsterdam-based computational chemistry software company. It was originally a spinout from the Vrije Universiteit. SCM supports and develops the ADF Modelling Suite, centred around the flagship program Amsterdam Density Functional (ADF), which was originally developed in the 1970s.

A standalone part of the modelling suite, named DFTB, is a fast-approximate Density Functional Theory approach for molecules and periodic systems.

What is DFTB?

The Density-Functional based Tight-Binding technique enables calculations on large systems for long timescales even on a desktop computer. Relatively accurate results are obtained at a fraction of the cost of full DFTB by reading in pre-calculated parameters (Slater–Koster files), using a minimal basis and only including nearest-neighbour interactions. Long-range interactions are described with empirical dispersion corrections and third-order corrections accurately handle charged systems.

The SCM implementation of the DFTB method can perform single point calculations, geometry optimizations, transition state searches, frequency calculations, and molecular dynamics. Molecules as well as periodic systems can be handled ensuring a smooth link with full DFT codes. It can be used as a stand-alone command line program, or from the graphical interface.

Objective

The objective of this work was to investigate the possibilities of improving the DFTB application performance on large systems distributed-memory machines. 

Method

Audit – An initial audit of the code took two months and focused on a density matrix purification technique in DFTB.  A set of performance metrics were used to assess the performance and identify any limiting issues, these metrics relate to computational scaling, load balance and communication. This method yielded a number of interesting results and identified an opportunity for improvement of its matrix multiplication kernel.

Report – The audit findings were delivered to the DFTB development group. The application was shown to have very good computational load balance, but it was recommended that some improvements should be made in the area of the MPI based communications code. Changes were expected to improve scalability on distributed-memory HPC platforms. In particular, it was suggested that it should be worthwhile to look into ways that communication could be conducted at the same time as computation.

Code improvements – For this project the advised enhancements were implemented as a proof of concept (POC) by the nAG team with direction from the SCM developers. The POC took four months to deliver and was tested on a large HPC machine. It clearly demonstrated that a more up-to-date version of the MPI Library had potential for improving the performance of the application, but unfortunately that was unavailable for testing on the development machine used at the time.

On-going – Even though there was not an explicit test with the newest versions of the MPI Library in the POC, the results were encouraging enough that the developers of the application decided to continue development of the improved code in the future. Further investigations will take place to look for performance scalability improvements when there is time.

Note: This work was carried out by nAG staff working under the remit of the Performance Optimisation and Productivity Centre of Excellence in Computing Applications (POP)

The post Audit of a Density Functional Theory application suggested a potentially improved communications scheme appeared first on nAG.

]]>
nAG Partner | DemandTec https://nag.com/case-studies/nag-partner-demandtec/ Thu, 06 Jul 2023 14:06:08 +0000 https://nag.com/?post_type=case-studies&p=3129 In today’s highly competitive retail market retailers need the most advanced and up to date software capabilities to achieve advantage over competitors and maximize profits.

The post <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG Partner | DemandTec appeared first on nAG.

]]>

nAG components add statistical advantage to global retailers

In today’s highly competitive retail market retailers need the most advanced and up to date software capabilities to achieve advantage over competitors and maximize profits. DemandTec provides the retail and consumer products industries with Consumer-Centric Merchandising software, which enables the planning, optimization and execution of merchandising and marketing programs based on a holistic understanding of consumer demand. No other company has as broad a merchandising planning and optimization software suite as DemandTec, nor does any other vendor have the large number of successful deployments with blue chip companies around the world.

A Different World

Retailers today face a different world than even five years ago. Fierce competition and the availability of massive amounts of data has driven even regional retailers to optimize supply chains, pricing and other critical aspects of their business. To meet these needs, DemandTec, Inc., the retail industry’s leading Consumer-Centric Merchandising software provider, recognized the need to deliver, to retail clients, modelling and analytical capabilities that would have seemed like science fiction just a few years earlier.

To realize its vision for a software architecture that could meet these demands, DemandTec knew that it needed an advanced analytics partner that could meet its current requirements and work with its science team to rapidly create new functionality as customer requirements evolved. After an intensive evaluation process DemandTec turned to nAG to supply the mathematical and statistical components needed in their software because of nAG’s longstanding reputation for quality and a track record of supplying code for dozens of ISV applications.

Code + Expertise = Results

In addition to nAG’s deserved reputation for accurate, robust and thoroughly documented code, DemandTec was particularly attracted to nAG’s extensive statistical functionality (300+ routines) including techniques such as Mixed Effect Regression, Correlation and Regression Analysis, Multivariate Methods and Time Series Analysis to name a few. Equally important for DemandTec was nAG’s staff of algorithmic experts and software engineers who were not only available to help the DemandTec team with integration issues but were also capable of creating entirely new methods for incorporation into DemandTec’s code.

Demand Management in Action

B&Q, the premier DIY retail chain in Europe utilizes DemandTec PriceTM, a web based software application, to more effectively manage their product pricing strategy. The software helps B&Q quickly understand how customers react to price changes to provide the best value to consumers while still meeting B&Q’s business goals.

Another blue chip company reaping the rewards of DemandTec’s Consumer-Centric Merchandising software is Fortune 50 food and drug retailer Safeway. DemandTec PriceTM and DemandTec PromotionTM support Safeway’s sales and marketing objectives and goals by enabling the company to determine and plan pricing and promotions based on the needs and wants of Safeway customers.

Partnership for Success

When it comes to creating cutting edge software to create retail and consumer products success, it is virtually impossible to assemble the customer knowledge, modelling science and algorithmic expertise within a single firm. DemandTec recognized this early on in its relationship with nAG. In the words of DemandTec CEO, Dan Fishback, “Partnering with nAG brought us world class software and algorithmic expertise that was a perfect complement to our customer, data and modelling knowledge and experience. nAG software and experts gave us the confidence to translate our ideas into market-leading software for our customers.

Rob Meyer, CEO of nAG, was involved in discussions with DemandTec from the beginning, adding “DemandTec quickly recognized how the market for demand management software was evolving and how their and our unique capabilities could combine to create unique solutions for customers. They have been an outstanding partner.

Learn More about DemandTec

DemandTec, an IBM Company, is a network of cloud apps and insights for more than 500 retailers and consumer products companies, providing common solutions to transact, interact, and collaborate on core merchandising and marketing activities. DemandTec’s services enable customers to achieve their sales volume, revenue, shopper loyalty, and profitability objectives. Approximately 16,000 retailer and manufacturer end-users on DemandTec collaborated on more than five million trade deals to date. DemandTec software and analytical services utilize a science-based platform to model and understand consumer behavior. DemandTec customers include leading retailers and consumer products companies such as Ahold USA, Best Buy, ConAgra Foods, Delhaize America, General Mills, The Home Depot, Monoprix, PETCO, Safeway, Sara Lee, Target, Walmart, and WHSmith.

The post <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG Partner | DemandTec appeared first on nAG.

]]>
HECToR dCSE Team Quadruples Speed and Enhances Scalability of Key Materials Science Code – CASTEP https://nag.com/case-studies/hector-dcse-team-quadruples-speed-and-enhances-scalability-of-key-materials-science-code-castep/ Wed, 05 Jul 2023 08:36:29 +0000 https://nag.com/?post_type=case-studies&p=2952 The objective of the dCSE project was to develop an improved, more scalable version of CASTEP – a software package which uses density functional theory with a plane wave basis set to calculate electronic properties of solids from first principles.

The post HECToR dCSE Team Quadruples Speed and Enhances Scalability of Key Materials Science Code – CASTEP appeared first on nAG.

]]>

CASTEP on HECToR

The UK’s previous national supercomputing facility, HECToR, has been heavily used by scientists needing capability supercomputing resources since its official launch in October 2007. A substantial part of the Research Councils’ six-year funding for this facility is devoted to the Computational Science and Engineering (CSE) support provided by nAG. An important part of the CSE Service is the distributed CSE (dCSE) programme which, through a lightweight peer review process, delivers dedicated multi-month performance and scalability development projects in response to proposals from the user community. The first dCSE project to complete has proven to be an excellent example of what can be achieved through dedicated CSE effort, with dramatic improvements in code performance and scalability which could potentially save millions of pounds and allow significant new science to be undertaken for the UK Car-Parrinello Consortium (UKCP).

Project Background

The objective of the dCSE project was to develop an improved, more scalable version of CASTEP – a software package which uses density functional theory with a plane wave basis set to calculate electronic properties of solids from first principles. The key task of the project was to implement band-parallelism in order to allow the code to scale to more than 1000 cores on HECToR. CASTEP is used on HECToR to model a range of materials or molecules at the atomic level. In particular scientists run CASTEP to obtain information about total energies, forces and stresses on an atomic system, as well as calculating optimum geometries, band structures, optical spectra, phonon spectra as well as molecular dynamics simulations. Dr Keith Refson from the Computational Materials Science Group at the Rutherford Appleton Laboratory was the Principal Investigator on the project, and Dr Martin Plummer from Daresbury Laboratory and Dr Matt Probert of The University of York were the Co-Investigators. nAG contracted Dr Phil Hasnip of the Department of Physics at the University of York to carry out the code development work in collaboration with both the wider CASTEP team and the nAG CSE team.

Project Results

The results of this work were excellent. The improved code has a speed-up factor of between 2 and 4 times the original and now scales to over 1000 cores against 256 previously.

The UKCP Chairman, Dr Matt Probert of the Department of Physics at the University York estimated that the CASTEP consortium was using around 10m Allocation Units (AU)s per annum on HECToR – a nominal cost of around £640k. Making the code 2-4 times more efficient could result in a saving of £320k-£480k per annum (a saving of around say £1.6m-£2.4m over the remaining life of HECToR); all for around 8 person months of effort!

Commenting on the massive return on investment, Dr Probert said 

 

“I guess it goes to show the value of centrally supporting key software packages, and that there is a considerable saving to be made due to scale of usage. Also the HECToR dCSE scheme is well worth supporting and continuing – and that the dCSE postdoc (Phil Hasnip in this case) was very good value for money!”

 

CASTEP user and PI on the project, Dr Keith Refson added 

 

“The performance and scaling gains achieved by the band-parallel CASTEP represent a very substantial advance in the efficiency of utilisation of CPU cycles on HECToR. This will not only result in a lower time and cost and more rapid turnaround for jobs already planned, but as intended will permit larger and more complex simulations, using more processors, which were not previously feasible.”

 

Dr Probert predicts that due to the speed and scaling gains resulting from CASTEP’s improvements the software can now be utilized for larger scale scientific work, i.e. bigger atomic systems for less wall-clock time and/or more simulations– within their existing budgets. In fact, several research projects have been waiting in anticipation of the planned release of the improved CASTEP in 2009.

Update

Since the writing of this case study, ARCHER has replaced HECToR as the UK’s national supercomputer.

About UKCP

The UKCP is an association of academic research groups collaborating on the first-principles computer simulation of condensed matter. Their aim is to use the creativity and strength of collaboration and shared expertise to produce outstanding science and simulation software.

About CASTEP

CASTEP is a software package which uses density functional theory with a plane wave basis set to calculate electronic properties of solids from first principles. CASTEP is a fully featured first principles code and as such its capabilities are numerous. Aiming to calculate any physical property of the system from first principles, the basic quantity is the total energy from which many other quantities are derived. For example the derivative of total energy with respect to atomic positions results in the forces and the derivative with respect to cell parameters gives stresses. These are then used to perform full geometry optimizations and possibly finite temperature molecular dynamics. Furthermore, symmetry and constraints (both internal and external to the cell) can be imposed in the calculations, either as defined by the user, or automatically using in-built symmetry detection.

The post HECToR dCSE Team Quadruples Speed and Enhances Scalability of Key Materials Science Code – CASTEP appeared first on nAG.

]]>
Performance Improvements on Suite of Computational Chemistry HPC Codes https://nag.com/case-studies/performance-improvements-on-suite-of-computational-chemistry-hpc-codes/ Tue, 04 Jul 2023 16:03:56 +0000 https://nag.com/?post_type=case-studies&p=2884 We performed a broad investigation of the ADF Modeling Suite including assessing the performance of new functionality, locating and fixing bottlenecks and testing multiple approaches on three of the most widely used applications

The post Performance Improvements on Suite of Computational Chemistry HPC Codes appeared first on nAG.

]]>

Project Background

The ADF Modeling Suite is a powerful computational chemistry package produced by SCM (Software for Chemistry and Materials), a company based in the Netherlands. It tackles chemistry, materials science and engineering challenges and is comprised of six applications.

We performed a broad investigation of the ADF Modeling Suite including assessing the performance of new functionality, locating and fixing bottlenecks and testing multiple approaches on three of the most widely used applications: ADF, BAND and DFTB. To analyse the performance of these applications, we used the tools Extrae, Paraver and Dimemas from Barcelona Supercomputing Center, along with Score-P and Scalasca.

ADF

ADF is a computational chemistry application which uses density functional theory calculations to predict the structure and reactivity of molecules. Sally Bridgwater, a Technical Consultant at nAG, assessed a particular calculation of interest to the developers which highlighted new functionality (medium-sized molecule with hybrid exchange-correlation functional). Initial investigations identified load imbalance as the main cause of inefficiency, resulting in idle time and the processors not being fully utilised. A secondary issue was the increase in time spent transferring data with increased core counts. ADF is parallelised with MPI and shared POSIX array buffers within a node. These shared buffers are not captured automatically by the profiling tools and so had to be manually instrumented using the Extrae API. The locking and unlocking of these shared arrays was found to exacerbate the load imbalance, leading to more idle time than initially thought.

A factor of two improvement was achieved

The long waiting times were found to be due to the load balancing algorithm not distributing work frequently enough and this was made worse by the imbalanced input used in this specific case. By distributing the work more frequently and tuning the chunk size of the work, Sally predicted that a factor of two saving in runtime could be achieved here. This gave SCM a clear idea of the potential improvement in performance.

SCM adapted their load balancing scheme to provide work more frequently by using a dedicated process to manage load balance, as well as initially dividing the work up into smaller chunks. SCM made these changes quickly and the estimated improvement of a factor of two was achieved.

Image: Timelines for the original (top) and version after improvements were made (bottom). This shows the large reduction in idle time (red) and halving of the total time taken.

Alexei Yakovlev, Software developer SCM, said:

“I can honestly say your analysis gave us a new insight into performance of one of the newer features available in ADF. What is more important, it clearly showed us the limitations of the current implementation and pointed us to the ways to improve it.”

BAND

The initial assessment of the BAND application by Jonathan Boyle, HPC Application Analyst at nAG, found profiling difficult due to the long runtimes and large amount of data produced for a moderately sized system. Initial findings, however, showed that load balance was the main bottleneck and that this was exacerbated by poor computational scalability.

Two more detailed assessments investigated sub-components of BAND. The first of these looked at computation of the overlap matrix. The performance observed ranged from reasonable to good, depending on the system being computed. Jonathan found that the main issue was reduced computational scaling, with contributions from low instructions scalability and IPC scalability. The routines responsible for the largest increases in exclusive time were identified for further investigation by the code developers.

The performance of the complex matrix-matrix multiplications was the focus of the second assessment, where each complex matrix is held using two shared memory real arrays, which are replicated on each compute node. This was found to have significant room for improvement and so a Proof of Concept study followed on from this work. The main bottlenecks identified were a reduction in IPC at scale and an increasing amount of time spent in MPI data transfer. Most computational work is within a dgemm call and hence this became a target for optimisation.

The Proof of Concept work that Jonathan implemented first tested out approaches to improve the performance and then implemented the ones that were beneficial in the source code.

The optimisations that were implemented are:

  • overlapping computation with communication
  • improved use of BLAS, which doubled the speed of computation
  • reorganising the algorithm to reduce the amount of data communicated via MPI.

The optimised subroutine showed four times speed up compared to the original code on eight 36-core compute nodes.

Image: Speedup plot showing the large improvement in scalability of the new subroutine.

DFTB

A case study of the work carried out on this application by Nick Dingle, Technical Consultant at nAG, can be read here. nAG was chosen to perform the following work due to our involvement in the POP Project.

The post Performance Improvements on Suite of Computational Chemistry HPC Codes appeared first on nAG.

]]>
Computational Cost Halved for Oil and Gas Seismic Code Through nAG HPC Expertise https://nag.com/case-studies/computational-cost-halved-for-oil-and-gas-seismic-code-through-nag-hpc-expertise-2/ Tue, 04 Jul 2023 15:30:31 +0000 https://nag.com/?post_type=case-studies&p=2860 NAG’s HPC team was chosen to perform the following work because of their experience in working with large HPC centers on specialist business focused projects. It was conducted within NAG’s involvement in the Performance Optimisation and Productivity Centre of Excellence.

The post Computational Cost Halved for Oil and Gas Seismic Code Through <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG HPC Expertise appeared first on nAG.

]]>

Code Background

nAG’s HPC team was chosen to perform the following work because of their experience in working with large HPC centers on specialist business focused projects. It was conducted within nAG’s involvement in the Performance Optimisation and Productivity Centre of Excellence.

Shearwater Reveal is a seismic processing code that does time and depth analysis for land and maritime applications. Its land processing tools cover all aspects of land processing, from refraction statistics to final time and depth imaging.

Application Analysis, Recommended Changes and Proof of Concept

The Shearwater Reveal application was assessed and the recommended changes were tested in a Proof of Concept (PoC). Wadud Miah, a HPC Application Analyst at nAG performed the work. During the assessment phase he identified reasonable load balance, good Amdahl’s efficiency and good computational efficiency. However, the OpenMP efficiency was low due to a high overhead.  

The cause of the high overhead was found to be an OpenMP critical region that protected file read/write operations from race conditions. In the PoC, Wadud modified the code such that the I/O was taken out of the OpenMP region, allowing the OpenMP critical region to then be removed. In addition, the OpenMP dynamic schedule was used. The parallel scalability graph below shows the changed PoC code with both the static and dynamic schedules.

The PoC code with the dynamic schedule only shows a performance gain at 18 and 24 threads. This is due to the increase in sequential execution caused by the memory allocation/deallocation needed to store the temporary data and the file I/O.

To investigate the potential performance gains, the memory allocation and file I/O were removed from the focus of analysis, even though these changes on their own led to an incorrect solution. The resulting scaling graph is shown below, with the linear and 80% of linear graphs scaled by the CPU frequency reductions.

Results

The Shearwater Reveal PoC showed a performance improvement as a result of these changes of up to 44%. The yellow line shows the parallel scaling with the I/O and memory allocation/deallocation removed and its performance is approaching linear scaling.

Additional recommendations were made by Wadud to aggregate the I/O into larger read/write sizes and to increase the re-use of data once it is read from disk.

Changes were made based on our analysis and recommendations. For a full-scale production run the computational cost was close to halved.

The post Computational Cost Halved for Oil and Gas Seismic Code Through <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG HPC Expertise appeared first on nAG.

]]>
nAG HPC Consulting Experts Enhance Computational Fluid Dynamics Application for Optimised Ship Design https://nag.com/case-studies/nag-hpc-consulting-experts-enhance-computational-fluid-dynamics-application-for-optimised-ship-design/ Tue, 04 Jul 2023 15:20:23 +0000 https://nag.com/?post_type=case-studies&p=2851 NAG, as HPC experts, were asked to investigate the optimal system configuration for the CFD based application to achieve good performance on the chosen HPC cluster platform.

The post <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG HPC Consulting Experts Enhance Computational Fluid Dynamics Application for Optimised Ship Design appeared first on nAG.

]]>

Project Background

Cetena is a study centre, based in Italy, that carries out a range of research and consulting work for the maritime industries. One of the software tools Cetena utilise as part of their virtual prototyping work is a computational fluid dynamics (CFD) application, known as HELYX-EcoMarine from ENGYS.

nAG, as HPC experts, were asked to investigate the optimal system configuration for the CFD based application to achieve good performance on the chosen HPC cluster platform.

Project Scope

Ship builders and operators, wanting to enhance their fleets, require ever more efficient and safer hull designs to cope with the demands of rough seas. Often, specialist consulting businesses are needed to provide expertise in fluid and environmental modelling. These subject specialists, such as Cetena, use advanced software for virtual modelling to optimize ship design. To assist with their virtual modelling work Cetena turn to application vendors, such as ENGYS, to provide the best software tools. nAG fits into this ecosystem of subject experts by providing the knowledge and experience of running a variety of numerically based software applications on large HPC clusters.

For this Cetena/ENGYS project nAG were funded by Fortissimo1 to assist the ISV partner to enhance its existing CFD package. nAG introduced a new solver for the analysis of added ship-hull hydrodynamics resistance problem, and incorporated a client-server component to facilitate the set-up, control and execution of large-scale simulations on remote HPC clusters via a desktop environment familiar to the end users.

nAG’s work included studying the overall architecture and software environment of the cluster, providing CSE support to the partners, simplifying the software deployment process, and establishing good practice on processing large distributed data sets and their remote visualisations.

Parallel 3D Remote Visualisation

Using open-source software, nAG investigated different approaches to supporting the visualisation of large distributed data sets on remote clusters. The combination of forward/reverse communication, with/without SSH tunnelling, and the use of different software rendering engines (and the use of GPU hardware wherever applicable) were studied. One efficient approach was taken forward and implemented by the ISV partner in their commercial CFD product.

Working cloud-based CFD solution

nAG’s work on this project has shown that a cloud-based solution can be achieved that delivers both efficiency (e.g. the 3D interactive visualisation of CFD models in real-time) and convenience (e.g. same desktop-based graphics user interface to set up and execute jobs, both locally and remotely). The cost-effectiveness of the cloud-based solution was also demonstrated. The project continues looking at how the software solution can be deployed in various cloud environments and how the deployment process can be streamlined to benefit the ISV partner.

1   This work was carried out as part of the collaborative European Fortissimo 2 project (www.fortissimo-project.eu) to promote the use of High Performance, cloud-based simulations in the engineering and manufacturing industries, mainly targeting SMEs and mid-cap companies.

The post <span class="nag-n-override" style="margin-left: 0 !important;"><i>n</i></span>AG HPC Consulting Experts Enhance Computational Fluid Dynamics Application for Optimised Ship Design appeared first on nAG.

]]>