Difference between revisions of "CyberShake Computational Estimates"

From SCECpedia
Jump to navigationJump to search
 
(6 intermediate revisions by the same user not shown)
Line 3: Line 3:
 
The UCERF 3 estimates assume that the number of ruptures increases from 15,000 to 350,000, but the number of rupture variations per rupture on average remains the same.
 
The UCERF 3 estimates assume that the number of ruptures increases from 15,000 to 350,000, but the number of rupture variations per rupture on average remains the same.
  
== Future SCEC data needs ==
+
The 0.5 Hz numbers are taken from Study 14.2.
  
These are estimates of CyberShake storage required on SCEC computers for upcoming runs.
+
The node-hours are estimates based on the XE6 and XK7 nodes on Blue Waters.
  
July-August 2013:  6.1 TB to archive (0.5 Hz, CVM-SI, 2 SGT versions, 286 sites, UCERF 2)
+
== 1.0 Hz, 3 component ==
Fall 2013:  12 TB (1 Hz, 286 sites, 2 combinations, UCERF 2)
 
Spring 2014:  45 TB (0.5 Hz, 286 sites, 1 combination, UCERF 3)
 
  
== UCERF 2 ==
+
SGTs: At 0.5 Hz, it requires 38 GPU node-hrs per component.
 +
  (38 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) x (20% more efficient due to more work per GPU) = 1460 node-hrs per site.
  
For UCERF 2 we have about 15,000 ruptures statewide, or 7000 ruptures / 415000 rupture variations per site.
+
23% of node hours to calculate SGTs.
  
=== 0.5 Hz (per site), 2 component ===
+
PP: At 0.5 Hz, it requires 41 CPU node-hrs per component.
 +
  (41 CPU node-hrs per components) x (3 components) x (25 times the rupture points) x (2 times the timesteps) x (20% more efficient due to rupture generator improvements) = 4920 node-hrs per site.
  
SGT data size:  40 GB
+
77% of node hours to calculate PP.
  
SGT SUs (CPU): 8k
+
Each site requires about 550,000 rupture variations (410,000 x 4/3 for rupture variations v3.3)
  
PP number of tasks: 415k
+
'''6380''' node-hours per 3-component site (181k core-hours)
  
PP SUs: 1k
+
'''1.82M''' node-hours for standard 3-component So Cal 286-site map (51.7M core-hours)
  
Seismogram data size: 11 GB
+
'''5.73M''' node-hours for increased density 3-component So Cal 898-site map (162M core-hours)
  
Total SUs: 9k
+
'''8.93M''' node-hours for statewide adaptive 3-component California 1400-site map (253M core-hours)
  
Total data: 51 GB
+
== 2.0 Hz ==
  
=== 1 Hz (per site), 3 component ===
+
SGTs: At 1.0 Hz, it requires 485 GPU node-hrs per component.
 +
  (485 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) = 23.3k node-hrs per site.
  
SGT data size: 1.9 TB (large increase due to resampling fault surfaces at 250 m)
+
PP: At 1.0 Hz, it requires 1640 CPU node-hrs per component.
 +
  (1640 CPU node-hrs per components) x (3 components) x (2 times the timesteps) = 9.8k node-hrs per site.
  
SGT SUs (CPU): 192k
+
'''33.1k''' node-hours per 3-component site (686k core-hours)
  
SGT SUs (GPU): 27k
+
'''9.47M''' node-hours for standard 3-component So Cal 286-site map (196M core-hours)
  
PP number of tasks: 415k
+
'''117M''' node-hours for increased density 3-component So Cal 3545-site map (2.4B core-hours)
  
PP SUs: 48k
+
'''46.3M''' node-hours for statewide adaptive 3-component California 1400-site map (960M core-hours)
 
 
Seismogram data size: 33 GB
 
 
 
Total SUs: 240k CPU / 75k GPU
 
 
 
Total data: 1.9 TB
 
 
 
=== 1.5 Hz (per site), 3 component ===
 
 
 
SGT data size:  2.8 TB
 
 
 
SGT SUs (CPU): 648k
 
 
 
SGT SUs (GPU): 93k
 
 
 
PP number of tasks: 415k
 
 
 
PP SUs: 72k
 
 
 
Seismogram data size: 50 GB
 
 
 
Total SUs: 720k CPU / 165k GPU
 
 
 
Total data: 2.8 TB
 
 
 
== UCERF 3 ==
 
 
 
For UCERF 3 we are expecting about 350000 ruptures statewide, or 163000 ruptures / 9.7 million rupture variations
 
 
 
=== 0.5 Hz (per site), 2 component ===
 
 
 
SGT data size:  930 GB
 
 
 
SGT SUs (CPU): 8k
 
 
 
PP number of tasks: 9.7M
 
 
 
PP SUs: 23k
 
 
 
Seismogram data size: 256 GB
 
 
 
Total SUs: 31k
 
 
 
Total data: 1.1 TB
 
 
 
=== 1 Hz (per site), 3 component ===
 
 
 
SGT data size: 44.5 TB
 
 
 
SGT SUs (CPU): 192k
 
 
 
SGT SUs (GPU): 27k
 
 
 
PP number of tasks: 9.7M
 
 
 
PP SUs: 1.1M
 
 
 
Seismogram data size: 768 GB
 
 
 
Total SUs: 240k CPU / 75k GPU
 
 
 
Total data: 45.2 TB
 
 
 
=== 1.5 Hz (per site), 3 component ===
 
 
 
SGT data size:  65.4 TB
 
 
 
SGT SUs (CPU): 648k
 
 
 
SGT SUs (GPU): 93k
 
 
 
PP number of tasks: 9.7M
 
 
 
PP SUs: 1.7M
 
 
 
Seismogram data size: 1.1 TB
 
 
 
Total SUs: 720k CPU / 165k GPU
 
 
 
Total data: 66.5 TB
 

Latest revision as of 15:52, 20 June 2014

We will describe or current best estimates for the CyberShake computational and data requirements as we progress in our simulation planning and testing. These estimates will help us identify which aspects of the CyberShake computational system needs to be optimized to work within our time and resource constraints.

The UCERF 3 estimates assume that the number of ruptures increases from 15,000 to 350,000, but the number of rupture variations per rupture on average remains the same.

The 0.5 Hz numbers are taken from Study 14.2.

The node-hours are estimates based on the XE6 and XK7 nodes on Blue Waters.

1.0 Hz, 3 component

SGTs: At 0.5 Hz, it requires 38 GPU node-hrs per component.

 (38 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) x (20% more efficient due to more work per GPU) = 1460 node-hrs per site.

23% of node hours to calculate SGTs.

PP: At 0.5 Hz, it requires 41 CPU node-hrs per component.

 (41 CPU node-hrs per components) x (3 components) x (25 times the rupture points) x (2 times the timesteps) x (20% more efficient due to rupture generator improvements) = 4920 node-hrs per site.

77% of node hours to calculate PP.

Each site requires about 550,000 rupture variations (410,000 x 4/3 for rupture variations v3.3)

6380 node-hours per 3-component site (181k core-hours)

1.82M node-hours for standard 3-component So Cal 286-site map (51.7M core-hours)

5.73M node-hours for increased density 3-component So Cal 898-site map (162M core-hours)

8.93M node-hours for statewide adaptive 3-component California 1400-site map (253M core-hours)

2.0 Hz

SGTs: At 1.0 Hz, it requires 485 GPU node-hrs per component.

 (485 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) = 23.3k node-hrs per site.

PP: At 1.0 Hz, it requires 1640 CPU node-hrs per component.

 (1640 CPU node-hrs per components) x (3 components) x (2 times the timesteps) = 9.8k node-hrs per site.

33.1k node-hours per 3-component site (686k core-hours)

9.47M node-hours for standard 3-component So Cal 286-site map (196M core-hours)

117M node-hours for increased density 3-component So Cal 3545-site map (2.4B core-hours)

46.3M node-hours for statewide adaptive 3-component California 1400-site map (960M core-hours)