1/3 hacluster units is in "Resource: res_manila_share_manila_share not yet configured"
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
OpenStack Manila-Ganesha Charm |
New
|
Undecided
|
Unassigned |
Bug Description
Looks like the manila-share resource have not been configured although it did get exposed over the relation between manila-ganesha and hacluster.
Observations:
* 1 of 3 units of manila-ganesha did not expose the resources to its hacluster unit;
* the leader hacluster unit have not even considered res_manila_
2020-08-05 17:42:52 DEBUG juju-log ha:42: Parsing cluster configuration using rid: ha:42, unit: manila-ganesha/0
...
2020-08-05 17:41:27 DEBUG juju-log hanode:2: Configuring Resources: {}
2020-08-05 17:43:07 DEBUG juju-log ha:42: Configuring Resources: {'res_ganesha_
...
2020-08-05 17:43:10 DEBUG ha-relation-changed ERROR: could not replace cib (rc=203)
2020-08-05 17:43:10 DEBUG ha-relation-changed INFO: offending xml: <diff format="2">
juju status:
manila-ganesha/0* active idle 11 10.5.0.17 Unit is ready
hacluster/0* active idle 10.5.0.17 Unit is ready and clustered
manila-ganesha/1 active idle 12 10.5.0.14 Unit is ready
hacluster/2 active idle 10.5.0.14 Unit is ready and clustered
manila-ganesha/2 active idle 13 10.5.0.18 Unit is ready
hacluster/1 waiting idle 10.5.0.18 Resource: res_manila_
ubuntu@
ubuntu@
clustered: "yes"
egress-subnets: 10.5.0.18/32
ingress-address: 10.5.0.18
private-address: 10.5.0.18
ubuntu@
Stack: corosync
Current DC: juju-dedf8c-
Last updated: Wed Aug 5 19:36:35 2020
Last change: Wed Aug 5 17:43:08 2020 by root via cibadmin on juju-dedf8c-
3 nodes configured
1 resource configured
Online: [ juju-dedf8c-
Full list of resources:
res_ganesha_
ubuntu@
corosync_bindiface: eth0
corosync_mcastport: "4440"
egress-subnets: 10.5.0.17/32
ingress-address: 10.5.0.17
json_colocations: '{"ganesha_
"manila_
json_delete_
json_resource_
migration-
interval=\"10s\" depth=\"0\""}'
json_resources: '{"res_
private-address: 10.5.0.17
ubuntu@
egress-subnets: 10.5.0.14/32
ingress-address: 10.5.0.14
private-address: 10.5.0.14
ubuntu@
corosync_bindiface: eth0
corosync_mcastport: "4440"
egress-subnets: 10.5.0.18/32
ingress-address: 10.5.0.18
json_colocations: '{"ganesha_
"manila_
json_delete_
json_groups: '{"grp_
json_resource_
migration-
interval=\"10s\" depth=\"0\"", "res_manila_
failure-
" meta migration-
json_resources: '{"res_
"systemd:
json_systemd_
private-address: 10.5.0.18
ubuntu@
clustered: "yes"
egress-subnets: 10.5.0.17/32
ingress-address: 10.5.0.17
private-address: 10.5.0.17
ubuntu@
clustered: "yes"
egress-subnets: 10.5.0.14/32
ingress-address: 10.5.0.14
private-address: 10.5.0.14
ubuntu@
clustered: "yes"
egress-subnets: 10.5.0.18/32
ingress-address: 10.5.0.18
private-address: 10.5.0.18
# cluster_connected got called at ~17:42:06
root@juju-
2020-08-05 17:42:06 INFO juju-log ha:42: Invoking reactive handler: reactive/
2020-08-05 17:42:21 INFO juju-log ceph:14: Invoking reactive handler: reactive/
2020-08-05 17:42:59 INFO juju-log manila-plugin:18: Invoking reactive handler: reactive/
root@juju-
Aug 05 17:36:19 juju-dedf8c-
Aug 05 17:36:21 juju-dedf8c-
Aug 05 17:36:22 juju-dedf8c-
Aug 05 17:42:02 juju-dedf8c-
Aug 05 17:42:10 juju-dedf8c-
# The first time ha.available was set for manila-ganesha/2
root@juju-
2020-08-05 17:43:33 DEBUG juju-log ha:42: tracer: set flag ha.available
ubuntu@
- Stdout: |
2020-08-05 17:39:37 DEBUG juju-log hanode:2: Configuring Resources: {}
2020-08-05 17:40:14 DEBUG juju-log hanode:2: Configuring Resources: {}
2020-08-05 17:40:50 DEBUG juju-log hanode:2: Configuring Resources: {}
2020-08-05 17:41:27 DEBUG juju-log hanode:2: Configuring Resources: {}
2020-08-05 17:43:07 DEBUG juju-log ha:42: Configuring Resources: {'res_ganesha_
UnitId: hacluster/0
- ReturnCode: 1
Stdout: ""
UnitId: hacluster/1
- ReturnCode: 1
Stdout: ""
UnitId: hacluster/2
1) hacluster/0 has never lost leadership during its lifetime (leader-elected got executed once and only on that unit);
2) the hacluster interface is container-scoped, therefore, the fact that manila-ganesha/1 has not exposed resource configuration does not matter;
3) the "Configuring Resources" log message appeared after the second ha-relation-changed got logged
* hacluster/0: 05 Aug 2020 17:37:47Z juju-unit executing running ha-relation-changed hook
* manila-ganesha/0:
2020-08-05 17:42:32 INFO juju-log ha:42: Invoking reactive handler: reactive/ manila_ ganesha. py:108: cluster_ connected
* hacluster/0:
05 Aug 2020 17:42:51Z juju-unit executing running ha-relation-changed hook
2020-08-05 17:43:07 DEBUG juju-log ha:42: Configuring Resources: # ...
Chronologically, hacluster/0 learned about the resources exposed by manila-ganesha/0 at 05 Aug 2020 17:42:51Z. There weren't any -changed events associated with that relation after that.
juju show-status-log hacluster/0 relation- joined hook relation- changed hook relation- joined hook relation- changed hook
Time Type Status Message
05 Aug 2020 17:36:14Z juju-unit executing running leader-elected hook
05 Aug 2020 17:36:20Z juju-unit executing running config-changed hook
05 Aug 2020 17:36:22Z workload maintenance Setting up corosync
05 Aug 2020 17:37:08Z juju-unit executing running start hook
05 Aug 2020 17:37:19Z workload active Unit is ready and clustered
05 Aug 2020 17:37:26Z juju-unit executing running ha-relation-joined hook
05 Aug 2020 17:37:47Z juju-unit executing running ha-relation-changed hook
05 Aug 2020 17:38:07Z juju-unit executing running hanode-
05 Aug 2020 17:38:27Z juju-unit executing running hanode-
05 Aug 2020 17:38:40Z workload blocked Insufficient peer units for ha cluster (require 3)
05 Aug 2020 17:38:48Z juju-unit executing running hanode-
05 Aug 2020 17:39:08Z juju-unit executing running hanode-
05 Aug 2020 17:41:39Z juju-unit idle
05 Aug 2020 17:42:51Z juju-unit executing running ha-relation-changed hook
05 Aug 2020 17:43:26Z juju-unit idle
juju show-status-log hacluster/1 --days 2
Time Type Status Message
05 Aug 2020 17:35:02Z workload waiting waiting for machine
05 Aug 2020 17:35:02Z juju-unit allocating
05 Aug 2020 17:35:02Z workload waiting installing agent
05 Aug 2020 17:35:06Z workload waiting agent initializing
05 Aug 2020 17:35:19Z workload maintenance installing charm software
05 Aug 2020 17:35:19Z juju-unit executing running install hook
05 Aug 2020 17:35:20Z workload maintenance Installin...