Pawsey Scheduled Maintenance (March)
Scheduled Maintenance Report for Pawsey Supercomputing Research Centre
Completed
Garrawarla has been returned to service.
Posted Mar 06, 2024 - 14:06 AWST
Verifying
Setonix has been returned to service.

Please note the minio client has been removed and the Lmod cache has been purged to support this change.

Scratch and Software has been upgraded to NEO 6.6-021 and global file locking has been re-instated. We will be looking at enforcing file limits in the near future.

The AMD GPU driver has been updated on all the GPU nodes to the version which comes with ROCm 5.5.3.

Garrawarla is currently being tested before returning to service.

If you have any questions, please e-mail help@pawsey.org.au.
Posted Mar 06, 2024 - 13:35 AWST
In progress
Scheduled maintenance is currently in progress. We will provide updates as necessary.
Posted Mar 05, 2024 - 07:00 AWST
Scheduled
Maintenance will be carried out on Setonix and Garrawarla on Tuesday the 5th of March to allow HPE to update the firmware of the scratch filesystem to restore the previous file locking configuration. HPE estimates the work will take two days, so the expected return to service is sometime on Thursday, 7th of March.

Please note after this work quota enforcement will be re-enabled (limiting users to 2 million files on scratch), cluster wide file locking with be re-enabled (flock) and the mc client will be removed from Setonix.

We will also be updating the GPU driver on all GPU nodes to the version which comes with ROCm 5.5.3. This will allow a new version of PyTorch container to be installed on Setonix

We appreciate your support and ask if you have any questions, please e-mail help@pawsey.org.au.
Posted Feb 27, 2024 - 09:51 AWST
This scheduled maintenance affected: Garrawarla (Garrawarla workq partition, Garrawarla gpuq partition, Garrawarla asvoq partition, Garrawarla copyq partition, Garrawarla login node, Slurm Controller (Garrawarla)), Lustre filesystems (/scratch filesystem (new), /software filesystem), and Setonix (Login nodes, Data-mover nodes, Slurm scheduler, Setonix work partition, Setonix debug partition, Setonix long partition, Setonix copy partition, Setonix askaprt partition, Setonix highmem partition, Setonix gpu partition, Setonix gpu high mem partition, Setonix gpu debug partition).