[CosmoMC + CosmoChord + ModeCode] Error in testing test_plan
Posted: August 11 2017
I managed to get {CosmoMC + CosmoChord} and {CosmoMC + CosmoChord + ModeCode} working (finally!) while just using these modules:
I have successfully carried out testing on test.ini for each of the two configurations. However, test_planck.ini yields errors. Before elaborating on that, I should point out that in the .bashrc file in my account on the cluster, I commented the line
because the cluster folks found out that this was the cause of the error outlined in the last post of this thread: http://cosmocoffee.info/viewtopic.php?t ... 502461a9f3. After that, I could successfully build the two configurations outlined above.
Now, when trying to get test_planck.ini to work, I either get one of the two errors (not sure what's the cause; am also in contact with cluster folks):
or
In fact, when I tried doing what was outlined here: http://cosmocoffee.info/viewtopic.php?p=5731 and also logged back out and in, I get the SIGILL error..
Please advise, thanks!
Code: Select all
Currently Loaded Modules:
1) GCCcore/6.3.0 3) GCC/6.3.0
2) binutils/2.27-GCCcore-6.3.0 4) OpenMPI/2.0.2-GCC-6.3.0
Code: Select all
source /projects/uoa00518/plc-2.0/bin/clik_profile.sh
because the cluster folks found out that this was the cause of the error outlined in the last post of this thread: http://cosmocoffee.info/viewtopic.php?t ... 502461a9f3. After that, I could successfully build the two configurations outlined above.
Now, when trying to get test_planck.ini to work, I either get one of the two errors (not sure what's the cause; am also in contact with cluster folks):
Code: Select all
Program received signal SIGILL: Illegal instruction.
Backtrace for this error:
Program received signal SIGILL: Illegal instruction.
Backtrace for this error:
#0 0x2b0f909d091f in ???
#1 0x50155c in ???
#2 0x403e5c in ???
#3 0x2b0f909bccdc in ???
#4 0x403eac in ???
#0 0x2b900db7e91f in ???
#1 0x50155c in ???
#2 0x403e5c in ???
#3 0x2b900db6acdc in ???
#4 0x403eac in ???
srun: error: compute-a1-066: task 0: Illegal instruction (core dumped)
srun: error: compute-a1-068: task 1: Illegal instruction (core dumped)
#0 0x2b1ab620091f in ???
#1 0x50155c in ???
#2 0x403e5c in ???
#3 0x2b1ab61eccdc in ???
#4 0x403eac in ???
#0 0x2b48f48d191f in ???
#1 0x50155c in ???
#2 0x403e5c in ???
#3 0x2b48f48bdcdc in ???
#4 0x403eac in ???
srun: error: compute-gpu-a1-002: task 3: Illegal instruction (core dumped)
srun: error: compute-gpu-a1-001: task 2: Illegal instruction (core dumped)
Code: Select all
Number of MPI processes: 4
file_root:test
NOTE: num_massive_neutrinos ignored, using specified hierarchy
NOTE: num_massive_neutrinos ignored, using specified hierarchy
Random seeds: 3286, 12940 rand_inst: 1
Random seeds: 4014, 12903 rand_inst: 4
NOTE: num_massive_neutrinos ignored, using specified hierarchy
NOTE: num_massive_neutrinos ignored, using specified hierarchy
Random seeds: 2911, 12974 rand_inst: 3
Random seeds: 3300, 13013 rand_inst: 2
compile with CLIK to use clik - see Makefile
MpiStop: 1
compile with CLIK to use clik - see Makefile
MpiStop: 2
compile with CLIK to use clik - see Makefile
MpiStop: 3
compile with CLIK to use clik - see Makefile
MpiStop: 0
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 1 in communicator MPI_COMM_WORLD
with errorcode 128.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 3 in communicator MPI_COMM_WORLD
with errorcode 128.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 2 in communicator MPI_COMM_WORLD
with errorcode 128.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
--------------------------------------------------------------------------
MPI_ABORT was invoked on rank 0 in communicator MPI_COMM_WORLD
with errorcode 128.
NOTE: invoking MPI_ABORT causes Open MPI to kill all MPI processes.
You may or may not see output from other processes, depending on
exactly when Open MPI kills them.
--------------------------------------------------------------------------
srun: error: compute-e1-001: task 0: Exited with exit code 128
srun: error: compute-e1-013: task 3: Exited with exit code 128
srun: error: compute-e1-002: task 1: Exited with exit code 128
srun: error: compute-e1-007: task 2: Exited with exit code 128
Please advise, thanks!