We've updated our Privacy Policy to make it clearer how we use your personal data. We use cookies to provide you with a better experience. You can read our Cookie Policy here.

Advertisement

The Future of Sustainable HPC

The Future of Sustainable HPC content piece image
Listen with
Speechify
0:00
Register for free to listen to this article
Thank you. Listen to this article using the player above.

Want to listen to this article for FREE?

Complete the form below to unlock access to ALL audio articles.

Read time: 3 minutes

It is an unavoidable fact that high-performance computing (HPC) is an energy-intensive environment. But what steps can we take to address this and achieve a more sustainable future for HPC in universities, research centers and other organizations?

Primary considerations include getting the most out of your power usage and considering renewable electricity sources. There are also options in offsetting the power you use within your organization to make your operations more sustainable and ‘green’. Organizations like Plan Vivo can help verify that the carbon offset program you choose to work with really is sustainable. The important thing to remember with offsetting is that it is best used as a tool to incentivize the reduction of resource consumption.

Offsetting helps to internalize the environmental cost of consuming power. Nevertheless for it to be effective the additional cost must be passed on to the people consuming that power. If users are currently being billed or allocated resource based on core hours / job run time they won’t see the additional power cost as clearly. Switching to a resource allocation scheme based on power consumption would be more effective. Tools such as the EAR energy management framework can be used to provide per-job energy consumption accounting through integration with Intel CPUs and SLURM accounting functionality.

When implementing a new build HPC system, it is important to understand what your workload is going to look like. Energy efficiency comes down to the level of utilization within the cluster. There are definitely architectures that are more energy-efficient. Generally, higher core count, lower clock speed processors tend to provide greater raw compute performance per watt, but you will need to have an application that will parallelize and is able to use all those hundreds of cores at once.

If your application doesn’t parallelize well, or if it needs higher frequency processors, then the best thing you can do is pick the right processor and the right number of them so you are not wasting power on CPU cycles that are not being used. When cycles are not being used, the CPUs should be configured to downclock to save power.

HPC managers will be assessed on how satisfied users are with their service, so many will artificially force all of the processors and nodes to run continually at 100 percent clock speed, so the processor won’t be put in a dormant state or be allowed to reduce its frequency. Ultimately, energy consumption just isn’t a major concern.

If customers come to us and want to improve energy efficiency based on their current estate, we would look at features being used in the scheduling software which can power off compute nodes, or at least put them into a dormant state if the processor supports that technology. We would check if these types of features are enabled and if they are making the most of them.

However, for some older clusters that do not support these features and generally provide much less power per watt performance than today’s technologies, it is often worth replacing a 200-node system that is 10 years old with something that is maybe 10 times smaller and provides just as much in terms of computing resource.

You can make a reasonable total cost of ownership (TCO) argument for ripping and replacing that entire old system. In some cases that will actually save money (and resources) over the next three to five years. Sometimes replacing what you have got is the best option, but I think the least invasive way and the first thing that we would look at with customers is whether they are being smart with their scheduling software and if there are benefits they can get in terms of reducing the power consumption of idle nodes.

Cloud bursting is another useful approach for sustainable HPC, particularly for instances when users have peak or infrequent workloads that don’t fit in their normal usage patterns. Using an automatic cloud resource for these workloads makes a lot of sense for temporary workloads. Rather than having 100 nodes on all the time, you may only need that capacity for four hours a day or once a week.

Large cloud providers can offer bigger benefits from economies of scale with power and energy efficiency using the latest cooling technologies and situating their data centers in more power efficient locations with greater renewable energy, like, for example, Iceland. This approach also helps to outsource concerns over environmental efficiencies for an organization.

Many universities have already voluntarily agreed to abide by the principles of the Paris Agreement for net zero by 2050. Some more ambitious institutions have committed to net zero by an earlier date. The high-performance computing departments of these universities will certainly become an important factor in these considerations in the near future.

Ultimately computing is burning through energy to produce computational results. You cannot get away from the fact that you need to use electricity to produce results, so the best thing you can do is to try and get the most computation out of every watt you use. That comes down to using your cluster to its maximum level, but then also making sure you are not wasting power.

By configuring the cluster in more environmentally sensitive ways, considering cloud options and giving people the conscious choice to do this will all help towards creating a more sustainable HPC industry in the future.

Mischa van Kesteren is pre-sales engineer and sustainability officer at OCF