-
Notifications
You must be signed in to change notification settings - Fork 137
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Issue with using pysph view #346
Comments
I am not clear as to what you are doing here. Are you using |
No, I was just using the command "pysph view dam_break_2d_output/" or "pysph view" without mpirun. Not sure why this error pops up. |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
I'm a beginner to pysph. I managed to install the package along with all the dependencies. Using the command "pysph run dam_break_2d" works for me and even prints the output data in the appropriate location. However, when I use pysph view dam_break_2d_output/, it throws up the following error:
It looks like orte_init failed for some reason; your parallel process is
likely to abort. There are many reasons that a parallel process can
fail during orte_init; some of which are due to configuration or
environment problems. This failure appears to be an internal failure;
here's some additional information (which may only be relevant to an
Open MPI developer):
setting topology failed
--> Returned value Error (-1) instead of ORTE_SUCCESS
It looks like orte_init failed for some reason; your parallel process is
likely to abort. There are many reasons that a parallel process can
fail during orte_init; some of which are due to configuration or
environment problems. This failure appears to be an internal failure;
here's some additional information (which may only be relevant to an
Open MPI developer):
orte_ess_init failed
--> Returned value Error (-1) instead of ORTE_SUCCESS
It looks like MPI_INIT failed for some reason; your parallel process is
likely to abort. There are many reasons that a parallel process can
fail during MPI_INIT; some of which are due to configuration or environment
problems. This failure appears to be an internal failure; here's some
additional information (which may only be relevant to an Open MPI
developer):
ompi_mpi_init: ompi_rte_init failed
--> Returned "Error" (-1) instead of "Success" (0)
*** An error occurred in MPI_Init_thread
*** on a NULL communicator
*** MPI_ERRORS_ARE_FATAL (processes in this communicator will now abort,
*** and potentially your MPI job)
[phani-B85M-DS3H:7967] Local abort before MPI_INIT completed completed successfully, but am not able to aggregate error messages, and not able to guarantee that all other processes were killed!
FYI, I'm using mpirun (Open MPI) 2.1.1.
Any direction to debug this issue will be greatly appreciated :)
The text was updated successfully, but these errors were encountered: