-
Notifications
You must be signed in to change notification settings - Fork 1
Performance Benchmarks Journal
Unless otherwise specified, all runs were relative to a Mac M1 platform, which is a 4+4 core platform
Danilo Lessa Bernardineli (BlockScience)
For varying N_sweep_samples
(N_t=2, N_s=2
):
- n=1: 0.09s (212= 4 state measurements, 44P/s)
- n=10: 0.52s (2102= 40 state measurements, 77P/s)
- n=100: 5.85s (21002= 400 state measurements, 68P/s)
- n=200: 12.03s (22002= 800 state measurements, 60P/s)
- n=400: 32.55s (24002= 1600 state measurements, 50P/s)
- n=600: error
- n=1000: error
For varying N_t
with N_sweep_samples=10
, N_s=2
:
- N_t=2: 0.51s (2102= 40 state measurements, 78P/s)
- N_t=20: 2.47s (2102= 400 state measurements, 162P/s)
- N_t=200: 42.72s (2102= 4000 state measurements, 93P/s)
For varying N_t
with N_sw=1
, N_s=10
- N_t=10: 0.7s (10110= 100 state measurements, 142P/s)
- N_t=100: 6.75s (101100 = 1,000 state measurements, 148P/s)
- N_t=1000: 283.0s (1011000 = 10,000 state measurements, 35P/s)
Last run: N_sw=20, N_s=2, N_t=100: 20.3s (220100= 100 state measurements, 197P/s)
As per the PSuU workplan, to be able to generate the measurements with an Exploratory coverage, we'll be required to be able to generate measurements across N_t=500
, N_s=3
and N_sw = 2048 * 209 = 428,032
, giving us a total of 2048*209*3*500 = 642,048,000
state measurements. Assuming a compute speed of 150P/s, we would require 49 days' worth of compute to generate the dataset. Having the Optimal coverage implies 3^9*2^2*50*209*5,000 = 4,113,747,000,000
state measurements, therefore requiring 317,418 days or a 6500x
coverage over the Exploratory one.
Assuming a target simulation duration of two hours for the Exploratory coverage, we would require a speed-up of 588x (or ~90,000 P/s
). As a hypothesis, we can assume we can have the following stacking performance multipliers by doing the following actions:
- Disabling
deepcopy
:10x
improvement - Model optimizations:
2x
improvement (20x
cumulative) - Simulation optimizations:
2x
improvement (40x
cumulative) - Running on a cloud VM:
20x
improvement (800x
cumulative)
For Optimal coverage, assuming a target simulation duration of 7 days, we would require a ~45,000x
speed-up. Therefore, we may require employing heuristics or iterative approaches.