Difference between revisions of "CyberShake Computational Estimates"

From SCECpedia
Jump to navigationJump to search
Line 3: Line 3:
 
The UCERF 3 estimates assume that the number of ruptures increases from 15,000 to 350,000, but the number of rupture variations per rupture on average remains the same.
 
The UCERF 3 estimates assume that the number of ruptures increases from 15,000 to 350,000, but the number of rupture variations per rupture on average remains the same.
  
== Future SCEC data needs ==
+
== 1.0 Hz ==
  
These are estimates of CyberShake storage required on SCEC computers for upcoming runs.
+
SGTs: At 0.5 Hz, it requires 35 GPU node-hrs per component.
 +
  (35 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) = 1680 node-hrs per site.
  
July-August 2013: 6.1 TB to archive (0.5 Hz, CVM-SI, 2 SGT versions, 286 sites, UCERF 2)
+
PP: At 0.5 Hz, it requires 30 CPU node-hrs per component.
Fall 2013:  12 TB (1 Hz, 286 sites, 2 combinations, UCERF 2)
+
  (30 CPU node-hrs per components) x (3 components) x (25 times the rupture points) x (2 times the timesteps) = 4500 node-hrs per site.
Spring 2014:  45 TB (0.5 Hz, 286 sites, 1 combination, UCERF 3)
 
  
== UCERF 2 ==
+
'''6180''' node-hours per 3-component site (200k core-hours)
  
For UCERF 2 we have about 15,000 ruptures statewide, or 7000 ruptures / 415000 rupture variations per site.
+
'''1.77M''' node-hours for standard 3-component So Cal 286-site map (56.6M core-hours)
  
=== 0.5 Hz (per site), 2 component ===
+
'''5.55M''' node-hours for increased density 3-component So Cal 898-site map (178M core-hours)
  
SGT data size:  40 GB
+
'''8.65M''' node-hours for statewide adaptive 3-component California 1400-site map (277M core-hours)
  
SGT SUs (CPU): 8k
+
== 2.0 Hz ==
  
PP number of tasks: 415k
+
SGTs: At 1.0 Hz, it requires 560 GPU node-hrs per component.
 +
  (560 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) = 27k node-hrs per site.
  
PP SUs: 1k
+
PP: At 1.0 Hz, it requires 1500 CPU node-hrs per component.
 +
  (1500 CPU node-hrs per components) x (3 components) x (2 times the timesteps) = 9k node-hrs per site.
  
Seismogram data size: 11 GB
+
'''36k''' node-hours per 3-component site (1.1M core-hours)
  
Total SUs: 9k
+
'''10.3M''' node-hours for standard 3-component So Cal 286-site map (328M core-hours)
  
Total data: 51 GB
+
'''127M''' node-hours for increased density 3-component So Cal 3545-site map (4.1B core-hours)
  
=== 1 Hz (per site), 3 component ===
+
'''50.2M''' node-hours for statewide adaptive 3-component California 1400-site map (1.6B core-hours)
 
 
SGT data size: 1.9 TB (large increase due to resampling fault surfaces at 250 m)
 
 
 
SGT SUs (CPU): 192k
 
 
 
SGT SUs (GPU): 27k
 
 
 
PP number of tasks: 415k
 
 
 
PP SUs: 48k
 
 
 
Seismogram data size: 33 GB
 
 
 
Total SUs: 240k CPU / 75k GPU
 
 
 
Total data: 1.9 TB
 
 
 
=== 1.5 Hz (per site), 3 component ===
 
 
 
SGT data size:  2.8 TB
 
 
 
SGT SUs (CPU): 648k
 
 
 
SGT SUs (GPU): 93k
 
 
 
PP number of tasks: 415k
 
 
 
PP SUs: 72k
 
 
 
Seismogram data size: 50 GB
 
 
 
Total SUs: 720k CPU / 165k GPU
 
 
 
Total data: 2.8 TB
 
 
 
== UCERF 3 ==
 
 
 
For UCERF 3 we are expecting about 350000 ruptures statewide, or 163000 ruptures / 9.7 million rupture variations.
 
 
 
=== 0.5 Hz (per site), 2 component ===
 
 
 
SGT data size:  930 GB
 
 
 
SGT SUs (CPU): 8k
 
 
 
PP number of tasks: 9.7M
 
 
 
PP SUs: 23k
 
 
 
Seismogram data size: 256 GB
 
 
 
Total SUs: 31k
 
 
 
Total data: 1.1 TB
 
 
 
=== 1 Hz (per site), 3 component ===
 
 
 
SGT data size: 44.5 TB
 
 
 
SGT SUs (CPU): 192k
 
 
 
SGT SUs (GPU): 27k
 
 
 
PP number of tasks: 9.7M
 
 
 
PP SUs: 1.1M
 
 
 
Seismogram data size: 768 GB
 
 
 
Total SUs: 240k CPU / 75k GPU
 
 
 
Total data: 45.2 TB
 
 
 
=== 1.5 Hz (per site), 3 component ===
 
 
 
SGT data size:  65.4 TB
 
 
 
SGT SUs (CPU): 648k
 
 
 
SGT SUs (GPU): 93k
 
 
 
PP number of tasks: 9.7M
 
 
 
PP SUs: 1.7M
 
 
 
Seismogram data size: 1.1 TB
 
 
 
Total SUs: 720k CPU / 165k GPU
 
 
 
Total data: 66.5 TB
 

Revision as of 20:26, 5 March 2014

We will describe or current best estimates for the CyberShake computational and data requirements as we progress in our simulation planning and testing. These estimates will help us identify which aspects of the CyberShake computational system needs to be optimized to work within our time and resource constraints.

The UCERF 3 estimates assume that the number of ruptures increases from 15,000 to 350,000, but the number of rupture variations per rupture on average remains the same.

1.0 Hz

SGTs: At 0.5 Hz, it requires 35 GPU node-hrs per component.

 (35 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) = 1680 node-hrs per site.

PP: At 0.5 Hz, it requires 30 CPU node-hrs per component.

 (30 CPU node-hrs per components) x (3 components) x (25 times the rupture points) x (2 times the timesteps) = 4500 node-hrs per site.

6180 node-hours per 3-component site (200k core-hours)

1.77M node-hours for standard 3-component So Cal 286-site map (56.6M core-hours)

5.55M node-hours for increased density 3-component So Cal 898-site map (178M core-hours)

8.65M node-hours for statewide adaptive 3-component California 1400-site map (277M core-hours)

2.0 Hz

SGTs: At 1.0 Hz, it requires 560 GPU node-hrs per component.

 (560 GPU node-hrs per component) x (3 components) x (8 times the gridpoints) x (2 times the timesteps) = 27k node-hrs per site.

PP: At 1.0 Hz, it requires 1500 CPU node-hrs per component.

 (1500 CPU node-hrs per components) x (3 components) x (2 times the timesteps) = 9k node-hrs per site.

36k node-hours per 3-component site (1.1M core-hours)

10.3M node-hours for standard 3-component So Cal 286-site map (328M core-hours)

127M node-hours for increased density 3-component So Cal 3545-site map (4.1B core-hours)

50.2M node-hours for statewide adaptive 3-component California 1400-site map (1.6B core-hours)