...
Section | ||||
---|---|---|---|---|
|
Past systems
For reference and record-keeping, here is the list of decommissioned systems that Pawsey operated in the past.
Magnus
Activity period: 2013 - 2022
Magnus was a Cray XC40 supercomputer, a massively parallel architecture consisting of 1,488 individual nodes that were connected via the Dragonfly network topology by a high-speed network, the HPC-optimised Aries interconnect. Each compute blade had a single Aries ASIC, which provides about 72Gbits/sec for each of the 4 nodes on the blade. All compute nodes in Magnus had the same architecture. Each included two Intel Xeon E5-2690 v3 (Haswell) 12-core CPUs for a total of 24 cores per node, providing a total of 35,712 cores for the entire system. Each node had 64 GB of DDR4 memory shared between 24 cores. Each core had 32KB instruction and data caches, and one 256KB L2 cache; 12 cores (per socket, or NUMA region) share one 30 MB L3 cache. In total, the system had 93 terabytes of memory.
Galaxy
Activity period: 2013 - 2022
Galaxy was a Cray XC30 supercomputer. It was only available for radio-astronomy-focused operations. In particular, it was used to support ASKAP and MWA, which are two of the Square Kilometre Array precursor projects currently underway in the northwest of Western Australia. For ASKAP, Galaxy acted as a real-time computer, allowing direct processing of data delivered to the Pawsey Centre from the Murchison Radio Observatory.
Zeus
Activity period: 2014 - 2022
Zeus was an HPE Linux cluster, containing different types of CPU nodes to support different types of computational workflows:
- For high-throughput workflows requiring large numbers of jobs that each use a modest number of cores:
An 80-node partition,workq
, was available. Each node had two Intel Xeon E5-2680 v4 2.4 GHz (Broadwell) 14-core CPUs and 128 GB of RAM. - For long-running workflows requiring several days to complete:
An 8-node partition,longq
was also available for computationally intensive jobs with wall times up to 4 days. Each node had 28 cores with 128 GB of RAM. - For large memory workflows requiring more than 128 GB of RAM:
A 6-node partition,higmemq
, was available for memory intensive computational jobs with 1TB of RAM each. Each node had 16 cores with 1TB of RAM. - For debugging and development work:
An 8-node partition,debugq
, was available for code development and prototyping work. Each node had 28 cores with 128 GB of RAM. - For data transfer jobs:
An 8-node partition,copyq
, was available for copying and transferring data. Each node contained two Intel Xeon E5-2650 (Sandy Bridge) 8-core CPUs and 64 GB of RAM.