[Users] how to speed up ET runs on cluster
Roland Haas
rhaas at illinois.edu
Tue Oct 23 10:05:39 CDT 2018
Hello Maria,
> I am using the Marshall University BigGreen cluster to run a
> parameter file with 5 Carpet time levels, on a 200^3 spatial domain,
> starting with d(x,y,z)_coarse = 1, and dtfac = 0.03125.
>
> The total necessary memory looks to be somewhere around 40 GB.
>
> My aim is to run it up to t=200, but I have a time limit on cluster
> of 72 hours.
You can run longer by checkpointing and recovering from the
checkpoints. In fact Cactus / Simfactory will do so automatically for
you if you ask for a runtime which is longer than the maxwalltime
setting of your machine.
> That obviously did not work, so I tried next to decrease the time
> refinement.
>
> For the same 48 procs, the run with bigger time step dtfac = 0.0625
> stopped ash t = 54.17 after 72 hours.
You usually cannot change the timestep and spatial resolution
independently since our runs are limited by the Courant–Friedrichs–Lewy
condition which links timestep and spatial resolution. So instead you should increase the spatial grid spacing (reduce the resolution) which would make the run go faster. Note that dtfac 0.03125 seems very small to me. Usually we end up with a dtfac of 0.25 - 0.4 depending on the timestepper used. Keep in mind that dtfac is the CFL factor for the coarsest level and that Carpet will automatically take smaller timesteps for the finer levels.
In your parfile, you would need to add:
IO::recover = "autoprobe"
IO::checkpoint_on_terminate = "yes"
Yours,
Roland
--
My email is as private as my paper mail. I therefore support encrypting
and signing email messages. Get my PGP key from http://pgp.mit.edu .
-------------- next part --------------
A non-text attachment was scrubbed...
Name: not available
Type: application/pgp-signature
Size: 833 bytes
Desc: OpenPGP digital signature
Url : http://lists.einsteintoolkit.org/pipermail/users/attachments/20181023/82936fb9/attachment.bin
More information about the Users
mailing list