component glance master scenario1 standalone fails tempestconf

Bug #1938558 reported by Marios Andreou
8
This bug affects 1 person
Affects Status Importance Assigned to Milestone
tripleo
Fix Released
Critical
Unassigned

Bug Description

At [1][2][3] the periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master fails after deployment and whilst running python-tempestconf to setup for tempest tests. The error occurs whilst trying to upload the cirros image to glance with trace like:

 2021-07-30 10:02:03.936 280634 INFO config_tempest.constants [-] Uploading image 'cirros-0.5.2-x86_64-disk.img' from '/home/zuul/tempest/etc/cirros-0.5.2-x86_64-disk.img'
 2021-07-30 10:02:04.009 280634 INFO tempest.lib.common.rest_client [req-634473e2-35d6-4860-a405-dc0e098c92a1 ] Request (main): 201 POST http://192.168.24.3:9292/v2/images 0.072s
 2021-07-30 10:02:10.659 280634 WARNING urllib3.connectionpool [-] Retrying (Retry(total=9, connect=None, read=None, redirect=5, status=None)) after connection broken by 'ProtocolError('Connection aborted.', BrokenPipeError(32, 'Broken pipe'))': /v2/images/023416d2-c796-4105-b58c-8dd3cb90c1f8/file
 2021-07-30 10:02:14.831 280634 INFO tempest.lib.common.rest_client [req-88b3b03e-6acb-4e55-baa7-65d3a5ebf90c ] Request (main): 500 PUT http://192.168.24.3:9292/v2/images/023416d2-c796-4105-b58c-8dd3cb90c1f8/file 10.821s
 2021-07-30 10:02:14.831 280634 CRITICAL tempest [-] Unhandled error: tempest.lib.exceptions.ServerFault: Got server fault
 Details: b'500 Internal Server Error\n\nThe server has either erred or is incapable of performing the requested operation.\n\n '
 2021-07-30 10:02:14.831 280634 ERROR tempest Traceback (most recent call last):
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/bin/discover-tempest-config", line 10, in <module>
 2021-07-30 10:02:14.831 280634 ERROR tempest sys.exit(main())
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/config_tempest/main.py", line 608, in main
 2021-07-30 10:02:14.831 280634 ERROR tempest retry_image=args.retry_image
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/config_tempest/main.py", line 547, in config_tempest
 2021-07-30 10:02:14.831 280634 ERROR tempest image.create_tempest_images(conf, retry_alt=retry_alt)
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/config_tempest/services/image.py", line 130, in create_tempest_images
 2021-07-30 10:02:14.831 280634 ERROR tempest retry_alt=retry_alt)
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/config_tempest/services/image.py", line 187, in find_or_upload_image
 2021-07-30 10:02:14.831 280634 ERROR tempest image = self._upload_image(image_name, image_dest)
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/config_tempest/services/image.py", line 243, in _upload_image
 2021-07-30 10:02:14.831 280634 ERROR tempest self.client.store_image_file(image['id'], data)
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/tempest/lib/services/image/v2/images_client.py", line 168, in store_image_file
 2021-07-30 10:02:14.831 280634 ERROR tempest body=data, chunked=True)
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/tempest/lib/common/rest_client.py", line 703, in request
 2021-07-30 10:02:14.831 280634 ERROR tempest self._error_checker(resp, resp_body)
 2021-07-30 10:02:14.831 280634 ERROR tempest File "/usr/lib/python3.6/site-packages/tempest/lib/common/rest_client.py", line 880, in _error_checker
 2021-07-30 10:02:14.831 280634 ERROR tempest message=message)
 2021-07-30 10:02:14.831 280634 ERROR tempest tempest.lib.exceptions.ServerFault: Got server fault
 2021-07-30 10:02:14.831 280634 ERROR tempest Details: b'500 Internal Server Error\n\nThe server has either erred or is incapable of performing the requested operation.\n\n '
 2021-07-30 10:02:14.831 280634 ERROR tempest 

This is blocking the master glance component promotions. The issue started happening on the 23rd (last successful run 22nd July @ [4])

[1] https://logserver.rdoproject.org/openstack-component-glance/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master/f778547/logs/undercloud/var/log/tempest/tempestconf.log.txt.gz
[2] https://logserver.rdoproject.org/openstack-component-glance/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master/a74e543/logs/undercloud/var/log/tempest/tempestconf.log.txt.gz
[3] https://logserver.rdoproject.org/openstack-component-glance/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master/86b8a46/logs/undercloud/var/log/tempest/tempestconf.log.txt.gz
[4] https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master

Revision history for this message
Marios Andreou (marios-b) wrote :

seeing errors in the glance container logs which seems to line up with the error seen in tempestconf timewise

https://logserver.rdoproject.org/openstack-component-glance/opendev.org/openstack/tripleo-ci/master/periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master/f778547/logs/undercloud/var/log/containers/glance/api.log.txt.gz

2021-07-30 10:02:10.477 180 DEBUG oslo_concurrency.processutils [-] 'blkid overlay -s UUID -o value' failed. Not Retrying. execute /usr/lib/python3.6/site-packages/oslo_concurrency/processutils.py:474
2021-07-30 10:02:10.477 180 DEBUG oslo.privsep.daemon [-] privsep: Exception during request[7003306d-4ba2-48e9-a117-88a327268641]: Unexpected error while running command.
Command: blkid overlay -s UUID -o value
Exit code: 2
Stdout: ''
Stderr: '' _process_cmd /usr/lib/python3.6/site-packages/oslo_privsep/daemon.py:482
Traceback (most recent call last):
  File "/usr/lib/python3.6/site-packages/oslo_privsep/daemon.py", line 477, in _process_cmd
    ret = func(*f_args, **f_kwargs)
  File "/usr/lib/python3.6/site-packages/oslo_privsep/priv_context.py", line 274, in _wrap
    return func(*args, **kwargs)
  File "/usr/lib/python3.6/site-packages/os_brick/privileged/rootwrap.py", line 197, in execute_root
    return custom_execute(*cmd, shell=False, run_as_root=False, **kwargs)
  File "/usr/lib/python3.6/site-packages/os_brick/privileged/rootwrap.py", line 146, in custom_execute
    on_completion=on_completion, *cmd, **kwargs)
  File "/usr/lib/python3.6/site-packages/oslo_concurrency/processutils.py", line 441, in execute
    cmd=sanitized_cmd)
oslo_concurrency.processutils.ProcessExecutionError: Unexpected error while running command.
Command: blkid overlay -s UUID -o value
Exit code: 2
Stdout: ''
Stderr: ''

Revision history for this message
Abhishek Kekane (abhishek-kekane) wrote :

(Thanks to Rajat for looking into this)

2 patches from cinder side are missing from the run
[1] python-cinderclient: https://review.opendev.org/c/openstack/python-cinderclient/+/783628
[2] cinder: https://review.opendev.org/c/openstack/cinder/+/783389

Revision history for this message
yatin (yatinkarel) wrote :

The actual error is:- TypeError: create() missing 1 required positional argument: 'instance_id'.
blkid overlay -s UUID -o value error is seen in successful jobs too, would be good to get that fixed too to avoid those errors.

<< 2 patches from cinder side are missing from the run
<< [1] python-cinderclient: https://review.opendev.org/c/openstack/python-cinderclient/+/783628
cinderclient-8.0.0 update in RDO waiting for https://review.opendev.org/q/topic:%22volumev2-removal%22+(status:open%20OR%20status:merged), some context can be found in https://review.rdoproject.org/r/c/rdoinfo/+/34695

<< [2] cinder: https://review.opendev.org/c/openstack/cinder/+/783389
cinder patch is there in the deployment, latest run have https://github.com/openstack/cinder/commit/b06a30.

The issue is caused with glance_store patch https://review.opendev.org/c/openstack/glance_store/+/782200. Which needs cinderclient-8.0.0 to get fixed and that's blocked as mentioned above. So we can pin glance_store to 2.6.0 in rdoinfo same as upper-constraints. Ideally it should have been pinned like other libraries, but need to check on history why it's not so or intentionally kept unpinned.
And when cinderclient-8.0.0 get's unblocked glance_store can also be unpinned back if needed.

Revision history for this message
Marios Andreou (marios-b) wrote :
Revision history for this message
yatin (yatinkarel) wrote :

@marios at https://github.com/redhat-openstack/rdoinfo/blob/master/tags/xena-uc.yml#L159, add source-branch: 2.6.0 as per u-c https://github.com/openstack/requirements/blob/master/upper-constraints.txt#L552

<< Ideally it should have been pinned like other libraries, but need to check on history why it's not so or intentionally kept unpinned.
I checked this and our automation didn't picked it as there is difference in project name in rdoinfo glance_store vs in upper-constraints glance-store. If storage team thinks need to follow u-c like other libraries, will work on fixing automation as this is the only exception i see. I think it should follow like other libraries and client but since it's not happening for long looking for confirmation. The only short time side effect will be RDO will build package for pinned tag, so going back for cases where there are new commits than last tag.

Revision history for this message
Marios Andreou (marios-b) wrote :

thanks @ykarel - I was a bit confused because I checked and 2.6.0 is the latest tag at https://github.com/openstack/glance_store/tags

but it looks like the offending patch at https://review.opendev.org/c/openstack/glance_store/+/782200 merged after that was made (14th July tag, 23rd july patch).

posted there https://review.rdoproject.org/r/c/rdoinfo/+/34869

Revision history for this message
Marios Andreou (marios-b) wrote :

thanks @ykarel

looks like the pin has helped. it merged Aug 06 4:36 PM [1] and then we had 2 green runs at [2]

        * openstack-component-glance 1 hr 9 mins 29 secs 2021-08-08 09:05:36 SUCCESS
        * openstack-component-glance 1 hr 19 mins 44 secs 2021-08-07 09:06:54 SUCCESS

[1] https://review.rdoproject.org/r/c/rdoinfo/+/34869/1#message-f7d33cddc3ee81b462f5b40043b88f26051fc710
[2] https://review.rdoproject.org/zuul/builds?job_name=periodic-tripleo-ci-centos-8-scenario001-standalone-glance-master

Revision history for this message
Marios Andreou (marios-b) wrote :

Just to close this out, after the pin at [1] to unblock us for this bug, the patch at [2] will ensure the pin is synchronized with upper constraints.

Moving the bug fix released please move back if you disagree and reach out to |ruck or |rover on oftc #tripleo or #oooq

[1] https://review.rdoproject.org/r/c/rdoinfo/+/34869
[2] https://review.rdoproject.org/r/c/rdo-infra/releng/+/34884

Changed in tripleo:
status: Triaged → Fix Released
To post a comment you must log in.
This report contains Public information  
Everyone can see this information.

Other bug subscribers

Remote bug watches

Bug watches keep track of this bug in other bug trackers.