<div dir="ltr"><div>As many of you know, restarting a simulation from checkpoints is often slower than it should be, and we have in the past implemented several performance improvements and work-arounds to address this. However, this is still an unsolved problem. The performance of accessing HDF5 datasets (i.e. opening the datasets and/or reading attributes) was often substantially slower than expected. The performance of reading the data themselves seems to be fine.</div><div><br></div><div>A few weeks ago, HDF5 1.10 was released. This features (optionally) a new file format that stores metadata (i.e. group directories and attributes) in a more efficient way. I always suspected that this should help with performance, but didn't run benchmarks at scale yet. Today, I see this email from Rob Latham <<a href="http://www.mcs.anl.gov/~robl/">http://www.mcs.anl.gov/~robl/</a>> on the HDF5 mailing list, which I copy below.</div><div><br></div><div>Essentially: HDF5 1.10 can provide a significant performance increase, and is worth a try.</div><div><br></div><div>If you try this at home, remember that HDF5 1.10 can create files that are backward compatible with 1.8. Obviously, if you do this, you will not see the performance benefits of the new file format. Please read the release notes to learn more about this, and about backward compatibility. Note that HDF5 1.10 also ships with a tool "h5format_convert" that converts from the 1.10 to the 1.8 format in case you have tools that can't read the new format. This conversion is very quick.</div><div><br></div><div>-erik</div><div><br></div><br><div class="gmail_quote">---------- Forwarded message ----------<br>From: <b class="gmail_sendername">Rob Latham</b> <span dir="ltr"><<a href="mailto:robl@mcs.anl.gov">robl@mcs.anl.gov</a>></span><br>Date: Tue, May 24, 2016 at 10:44 AM<br>Subject: [Hdf-forum] hdf-1.10 and collective metadata success story<br>To: <a href="mailto:hdf-forum@lists.hdfgroup.org">hdf-forum@lists.hdfgroup.org</a><br><br><br>Since I've spent oh, only the last 8 years beating you up a bit about HDF5 metadata reads at scale, it is only fair that I share with you some good news. Thanks for implementing a scalable metadata read. It's already paying off.<br>
<br>
There was a user trying to read a large dataset on our Blue Gene. We got him using 1.10, and presto:<br>
<br>
Here's a quote from the user:<br>
<br>
Yes, I can now restore data (with hdf5 1.10). I did a large test simulation on cetus (with fewer nodes than I'd usually run, so a lot of data per node), and then followed up by restoring a real 1024-node simulation on mira: this read 6 "field files" of 3.8 GB each and 2 "particle files" of 396 GB and 2 "particle files" of 69 GB in 442 seconds. Previously it was taking on the order of an hour to read in a 4 GB field file.<br>
<br>
==rob<br>
<br>
_______________________________________________<br>
Hdf-forum is for HDF software users discussion.<br>
<a href="mailto:Hdf-forum@lists.hdfgroup.org" target="_blank">Hdf-forum@lists.hdfgroup.org</a><br>
<a href="http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org" rel="noreferrer" target="_blank">http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org</a><br>
Twitter: <a href="https://twitter.com/hdf5" rel="noreferrer" target="_blank">https://twitter.com/hdf5</a><br>
</div><br><br clear="all"><div><br></div>-- <br><div class="gmail_signature">Erik Schnetter <<a href="mailto:schnetter@gmail.com" target="_blank">schnetter@gmail.com</a>> <a href="http://www.perimeterinstitute.ca/personal/eschnetter/" target="_blank">http://www.perimeterinstitute.ca/personal/eschnetter/</a><br></div>
</div>