oslo cache mempool issues with python3
Affects | Status | Importance | Assigned to | Milestone | |
---|---|---|---|---|---|
Ubuntu Cloud Archive |
Fix Released
|
High
|
Unassigned | ||
Rocky |
Fix Released
|
High
|
Unassigned | ||
Stein |
Fix Released
|
High
|
Unassigned | ||
oslo.cache |
Fix Released
|
High
|
Herve Beraud | ||
python-oslo.cache (Ubuntu) |
Fix Released
|
High
|
Unassigned | ||
Cosmic |
Fix Committed
|
High
|
Unassigned | ||
Disco |
Fix Released
|
High
|
Unassigned |
Bug Description
nova conductor running on a rhel8 host inside f28 based containers hits the following error:
2019-01-17 13:59:37.049 46 DEBUG oslo_concurrenc
2019-01-17 13:59:37.050 46 DEBUG oslo_concurrenc
2019-01-17 13:59:37.060 46 DEBUG oslo_db.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
2019-01-17 13:59:37.096 46 ERROR oslo_messaging.
nova.conf has:
()[nova@standalone /]$ grep -v ^# /etc/nova/nova.conf |grep '[a-zA-Z]' |grep -i mem
backend=
memcache_
memcached_
memcache seems to be up:
[root@standalone ~]# podman top memcached
USER PID PPID %CPU ELAPSED TTY TIME COMMAND
memcached 1 0 0.000 2h2m27.45050385s ? 0s dumb-init --single-child -- /bin/bash -c source /etc/sysconfig/
memcached 8 1 0.000 2h2m27.450722828s ? 0s /bin/bash -c source /etc/sysconfig/
memcached 9 8 0.000 2h2m27.450781466s ? 0s /usr/bin/memcached -p 11211 -u memcached -m 11970 -c 8192 -v -l 192.168.24.2 -U 0 -X -t 8 >> /var/log/
Workaround:
disable caching in nova
crudini --set /var/lib/
podman restart nova_conductor
Versions inside the f28 nova-conductor container:
python-
python3-
python3-
python3-
python-
puppet-
python3-
python3-
python3-
python-
python-
python3-
python3-
python-
python3-
python3-
python3-
python-
python-
python3-
python-
python3-
python3-
python3-
python3-
python-
python3-
python3-
python-
python-
python3-
puppet-
-------
Ubuntu SRU details
------------------
[Impact]
See description above.
[Test Case]
1. Deploy an HA Rocky cloud using the openstack-
2. Configure a simple network, upload a bionic image.
3. Try to start an instance via the API. It will stay in build.
4. Try to use the openstack cli to read availability zones. It will
return an error.
For more info see bug 1823740
[Regression Potential]
The regression potential is low. This is a minimal fix that has successfully been reviewed upstream and passed all upstream gate tests. It has already landed in upstream master branch and Ubuntu Disco and received 3 +1's and zuul tests +1 on stable/rocky gerrit reviews.
Changed in oslo.cache: | |
status: | New → Confirmed |
importance: | Undecided → High |
Changed in python-oslo.cache (Ubuntu Cosmic): | |
status: | New → Triaged |
Changed in python-oslo.cache (Ubuntu Disco): | |
status: | New → Triaged |
Changed in python-oslo.cache (Ubuntu Cosmic): | |
importance: | Undecided → High |
Changed in python-oslo.cache (Ubuntu Disco): | |
importance: | Undecided → High |
Changed in python-oslo.cache (Ubuntu Disco): | |
status: | Triaged → Fix Released |
tags: | added: cdo-qa cdo-release-blocker foundations-engine |
Changed in python-oslo.cache (Ubuntu Disco): | |
status: | Fix Released → In Progress |
status: | In Progress → Fix Released |
description: | updated |
Some debug informations to move a little bit forward on this topic...
The exception is raised inside the nova-conductor container.
The problem seems to occur when the cache is enable and the backend used is oslo.cache.
Config in use in /var/lib/ config- data/puppet- generated/ nova/etc/ nova/nova. conf : oslo_cache. memcache_ pool backend= True
backend=
enabled=True
debug_cache_
Nova services seems to be unregister: ------- +------ +------ +------ --+---- ---+--- ------- --+---- ------- ------+ ------- ------+ │·| Id | Binary | Host | Zone | Status | State | Updated_at | Disabled Reason | Forced down | │· ------- +------ +------ +------ --+---- ---+--- ------- --+---- ------- ------+ ------- ------+ │·+---- +------ --+---- --+---- --+---- ----+-- -----+- ------- ----+-- ------- ------- -+----- ------- -+
$ openstack compute service list
+----+-
+----+-
Else when the cache is disable the previous command return results ------- ------- ---+--- ------- ------- -+----- -----+- ------- -+----- --+---- ------- ------- ------- ---+ ------- ------- ---+--- ------- ------- -+----- -----+- ------- -+----- --+---- ------- ------- ------- ---+ 24T16:07: 47.000000 | 24T16:07: 41.000000 | 24T16:07: 43.000000 | 24T16:07: 47.000000 | ------- ------- ---+--- ------- ------- -+----- -----+- ------- -+----- --+---- ------- ------- ------- ---+
[stack@win1 ~]$ openstack compute service list
+----+-
| ID | Binary | Host | Zone | Status | State | Updated At |
+----+-
| 1 | nova-conductor | win1.localdomain | internal | enabled | up | 2019-01-
| 2 | nova-consoleauth | win1.localdomain | internal | enabled | up | 2019-01-
| 3 | nova-scheduler | win1.localdomain | internal | enabled | up | 2019-01-
| 8 | nova-compute | win1.localdomain | nova | enabled | up | 2019-01-
+----+-
Same problem when we use `nova service-list`