Difference between revisions of "MPI on virtual nodes"
Jump to navigation
Jump to search
(thanks Shane!) |
|||
(3 intermediate revisions by the same user not shown) | |||
Line 1: | Line 1: | ||
+ | =OBSOLETE= | ||
− | To run MPI on virtual nodes one has to specify the interface | + | |
+ | To run MPI on virtual nodes one has to specify the interface. | ||
The argument (to mpirun) needed is: | The argument (to mpirun) needed is: | ||
<code>--mca btl_tcp_if_include tun0</code> | <code>--mca btl_tcp_if_include tun0</code> | ||
+ | |||
+ | Alternatively one may specify the network to be used for communication to cover both real and virtual node networks: | ||
+ | |||
+ | <code>--mca btl_tcp_if_include 192.168.100.0/23</code> | ||
Source for more knowledge: | Source for more knowledge: | ||
http://stackoverflow.com/questions/10466119/mpi-send-stops-working-after-mpi-barrier/10473106#10473106 | http://stackoverflow.com/questions/10466119/mpi-send-stops-working-after-mpi-barrier/10473106#10473106 |
Latest revision as of 05:21, 10 January 2019
OBSOLETE
To run MPI on virtual nodes one has to specify the interface. The argument (to mpirun) needed is:
--mca btl_tcp_if_include tun0
Alternatively one may specify the network to be used for communication to cover both real and virtual node networks:
--mca btl_tcp_if_include 192.168.100.0/23
Source for more knowledge: http://stackoverflow.com/questions/10466119/mpi-send-stops-working-after-mpi-barrier/10473106#10473106