I cannot reproduce this issue manually. Here is the log. We can see that the node gets offline status is set within one second. So, I suppose there is some race condition. I would suggest to check for corosync process status on the first node and wait while corosync finally terminates and check pcs nodes status only after that. It would also be really useful to have timestamps for logs steps to be able to debug things.
root@node-6:~# killall -TERM corosync; date +%s.%N
1484752636.103218131
@Dmitry
I cannot reproduce this issue manually. Here is the log. We can see that the node gets offline status is set within one second. So, I suppose there is some race condition. I would suggest to check for corosync process status on the first node and wait while corosync finally terminates and check pcs nodes status only after that. It would also be really useful to have timestamps for logs steps to be able to debug things. 103218131
root@node-6:~# killall -TERM corosync; date +%s.%N
1484752636.
root@node-7 test.domain. local node-6. test.domain. local node-7. test.domain. local 759762323 test.domain. local node-7. test.domain. local test.domain. local 373758520
Pacemaker Nodes:
Online: node-5.
Standby:
Offline:
Pacemaker Remote Nodes:
Online:
Standby:
Offline:
1484752636.
Pacemaker Nodes:
Online: node-5.
Standby:
Offline: node-6.
Pacemaker Remote Nodes:
Online:
Standby:
Offline:
1484752637.