[ET Trac] [Einstein Toolkit] #180: Reduce disk space used by checkpoints
Einstein Toolkit
trac-noreply at einsteintoolkit.org
Fri Dec 31 07:36:02 CST 2010
#180: Reduce disk space used by checkpoints
--------------------------+-------------------------------------------------
Reporter: hinder | Owner: mthomas
Type: enhancement | Status: new
Priority: major | Milestone:
Component: SimFactory | Version:
Resolution: | Keywords:
--------------------------+-------------------------------------------------
Comment (by eschnett):
SimFactory's philosophy is that each restart is independent of the other
restarts, like a snapshop of the simulation in time. The hard-linking
mechanism was introduced to prevent Cactus from (accidentally?) deleting
data from old restarts.
One problem in long-running simulations is that (a) one may accidentally
delete too many checkpoints, so that the simulation becomes unusable.
Another problem is that, if an error in the simulation is detected and if
all old checkpoints have been deleted, it is impossible to "go back in
time" (using checkpoints left every so many time steps).
My current idea is to archive old restarts, and then delete them (or parts
of them) locally. In this way, no data have to be deleted. This depends on
the availability of archival storage, but this should be present -- we
mostly have not been using it because this is somewhat awkward: it takes a
long time, different systems have different archives, and it is cumbersome
to find out which files have been archived. However, it should in
principle be straightforward to automatically copy older restarts into an
archive.
--
Ticket URL: <https://trac.einsteintoolkit.org/ticket/180#comment:1>
Einstein Toolkit <http://einsteintoolkit.org>
The Einstein Toolkit
More information about the Trac
mailing list