[ET Trac] [Einstein Toolkit] #1283: Missing data in HDF5 files
Einstein Toolkit
trac-noreply at einsteintoolkit.org
Mon Mar 11 12:02:04 CDT 2013
#1283: Missing data in HDF5 files
---------------------+------------------------------------------------------
Reporter: hinder | Owner:
Type: defect | Status: new
Priority: major | Milestone:
Component: Cactus | Version:
Resolution: | Keywords:
---------------------+------------------------------------------------------
Comment (by eschnett):
I would not use "tmp" to indicate files are currently incomplete. What
about "incomplete"? Or "$$$"?
One either needs to store the data belonging to a simulation, or re-run
the simulation if there is a problem. If archives can't handle checkpoint
files, and if local data disks aren't large enough to keep checkpoint
files, then re-running is the correct response. People need to be aware of
this, and in this mode, all but the last (verified to be safely usable)
checkpoint files should be deleted, probably semi-automatically.
If a data disk has 300 Tbyte, then I would expect an archive to have 3,000
Tbyte. Alternatively, one could design a system that has 100 Tbyte data
disk and 1,000 Tbyte usable archive space (10 Gbit interconnect?).
Currently, since no one uses archive except to stash data they don't
intend to look at any more, the archive is difficult to use, and there is
a large pressure to increase data disks, further reducing the need to use
an archive, and thus reducing pressure to make it easier to use.
I meant checkpoint files. But, thinking about this, it could equally apply
to restarts, or at least large output files in restarts.
--
Ticket URL: <https://trac.einsteintoolkit.org/ticket/1283#comment:12>
Einstein Toolkit <http://einsteintoolkit.org>
The Einstein Toolkit
More information about the Trac
mailing list