[Users] Cactus Core Usage on HPC Cluster
Erik Schnetter
schnetter at cct.lsu.edu
Sat Feb 4 13:30:39 CST 2017
Gwyneth
This question is unrelated to Cactus, and only concerns the MPI
implementation you are using and the queuing system. Since you already
received advice regarding how to call mpirun, you should probably ask
that person for help.
A few questions:
- Is there an example MPI program that you can run to test this?
- Which MPI implementation did you use when building Cactus? Did you
check that this worked correctly? How?
- Can you point us to documentation regarding PBS and MPI for your cluster?
To be able to actually help, we will need more details. Consider
opening a help ticket and posting the details there, which would
include your option list, submit script, run script, relevant output,
etc.
-erik
On Sat, Feb 4, 2017 at 2:13 PM, Gwyneth Allwright <allgwy001 at myuct.ac.za> wrote:
> Hi All,
>
> I'm trying to get the Einstein Toolkit installed on an HPC cluster running
> SLES. The trouble is that Cactus tries to use all the available processors
> even when I specify a smaller number (by setting ppn in my PBS script).
>
> As a test, I tried running the compilation with a parameter file that
> required about 5 GB of RAM. In my PBS script, I set nodes=1 and ppn=3, and
> then ran using openmpi-1.10.1:
>
> mpirun -hostfile $PBS_NODEFILE <ET exe> <parameter file>
>
> This resulted in the simulation running on all 24 available processors, even
> though I'd only requested 3. Since PBS and MPI are integrated, I was told
> that using -np with mpirun wouldn't help.
>
> Does anyone know how to address this issue?
>
> Thanks,
>
> Gwyneth
>
>
> _______________________________________________
> Users mailing list
> Users at einsteintoolkit.org
> http://lists.einsteintoolkit.org/mailman/listinfo/users
>
--
Erik Schnetter <schnetter at cct.lsu.edu>
http://www.perimeterinstitute.ca/personal/eschnetter/
More information about the Users
mailing list